Recent breakthroughs in natural language processing show that attention mechanism in Transformer networks, trained via masked-token prediction, enables models to capture the semantic context of the tokens and internalize the grammar of language. While the application of Transformers to communication systems is a burgeoning field, the notion of context within physical waveforms remains under-explored. This paper addresses that gap by re-examining inter-symbol contribution (ISC) caused by pulse-shaping overlap. Rather than treating ISC as a nuisance, we view it as a deterministic source of contextual information embedded in oversampled complex baseband signals. We propose Masked Symbol Modeling (MSM), a framework for the physical (PHY) layer inspired by Bidirectional Encoder Representations from Transformers methodology. In MSM, a subset of symbol aligned samples is randomly masked, and a Transformer predicts the missing symbol identifiers using the surrounding "in-between" samples. Through this objective, the model learns the latent syntax of complex baseband waveforms. We illustrate MSM's potential by applying it to the task of demodulating signals corrupted by impulsive noise, where the model infers corrupted segments by leveraging the learned context. Our results suggest a path toward receivers that interpret, rather than merely detect communication signals, opening new avenues for context-aware PHY layer design.
翻译:自然语言处理领域的最新突破表明,通过掩码标记预测训练的Transformer网络中的注意力机制,能够使模型捕捉标记的语义上下文并内化语言的语法结构。尽管将Transformer应用于通信系统是一个新兴领域,但物理波形中的上下文概念仍未被充分探索。本文通过重新审视由脉冲成形重叠引起的符号间贡献来填补这一空白。我们不再将符号间贡献视为干扰,而是将其视为嵌入在过采样复基带信号中的确定性上下文信息源。受Transformer双向编码器表征方法的启发,我们提出了掩码符号建模这一面向物理层的框架。在MSM中,一部分符号对齐的样本被随机掩码,Transformer利用周围的“中间”样本预测缺失的符号标识符。通过这一目标,模型能够学习复基带波形的潜在语法结构。我们通过将MSM应用于脉冲噪声干扰下的信号解调任务来展示其潜力,在该任务中模型利用学习到的上下文推断受损信号段。研究结果表明,这为开发能够解释而不仅仅是检测通信信号的接收机提供了可能路径,为上下文感知的物理层设计开辟了新途径。