Recent work has revealed that state space models (SSMs), while efficient for long-sequence processing, are fundamentally limited in their ability to represent formal languages-particularly due to time-invariant and real-valued recurrence structures. In this work, we draw inspiration from adaptive and structured dynamics observed in biological neural systems and introduce the Adaptive Unitary State Space Model (AUSSM), a novel class of SSMs that leverages skew-symmetric, input-dependent recurrence to achieve unitary evolution and high expressive power. Using algebraic automata theory, we prove that AUSSM can perform modulo counting and simulate solvable group automata at finite precision, enabling AUSSM to model a broad class of regular languages out of reach for other SSM architectures. To overcome the practical inefficiencies of adaptive recurrence, we develop a separable convolution formulation and a CUDA implementation that enables scalable parallel training. Empirically, we show that AUSSM and its hybrid variant-interleaved with Mamba-outperform prior SSMs on formal algorithmic tasks such as parity and modular arithmetic, and achieve competent performance on real-world long time-series classification benchmarks. Our results demonstrate that adaptive unitary recurrence provides a powerful and efficient inductive bias for both symbolic and continuous sequence modeling. The code is available at https://github.com/arjunkaruvally/AUSSM
翻译:近期研究表明,态空间模型(SSMs)虽然在长序列处理中效率较高,但在表示形式语言方面存在根本性限制——这主要源于其时不变且为实值的递归结构。受生物神经系统中观察到的自适应与结构化动力学启发,本文提出自适应酉态空间模型(AUSSM),这是一类新型SSM,它利用斜对称且输入依赖的递归机制实现酉演化并具备高表达力。基于代数自动机理论,我们证明AUSSM能在有限精度下执行模运算计数并模拟可解群自动机,从而使其能够建模其他SSM架构无法覆盖的大范围正则语言类。为克服自适应递归在实践中的效率瓶颈,我们开发了可分离卷积形式及CUDA实现方案,支持可扩展的并行训练。实验表明,AUSSM及其与Mamba交错组合的混合变体,在奇偶校验与模运算等形式算法任务上超越现有SSM,并在现实世界长时序分类基准测试中取得优异性能。我们的研究证明,自适应酉递归机制为符号与连续序列建模提供了强大且高效的归纳偏置。代码已发布于https://github.com/arjunkaruvally/AUSSM