Recurrent neural networks (RNNs) are a cornerstone of sequence modeling across various scientific and industrial applications. Owing to their versatility, numerous RNN variants have been proposed over the past decade, aiming to improve the modeling of long-term dependencies and to address challenges such as vanishing and exploding gradients. However, no central library is available to test these variations, and reimplementing diverse architectures can be time-consuming and error-prone, limiting reproducibility and exploration. Here, we introduce three open-source libraries in Julia and Python that centralize numerous recurrent cell implementations and higher-level recurrent architectures. torchrecurrent, RecurrentLayers.jl, and LuxRecurrentLayers.jl offer a consistent framework for constructing and extending RNN models, providing built-in mechanisms for customization and experimentation. All packages are available under the MIT license and actively maintained on GitHub.
翻译:循环神经网络(RNNs)是序列建模的基石,广泛应用于各类科学与工业领域。由于其通用性,过去十年间涌现了大量旨在改善长程依赖建模、解决梯度消失与爆炸等挑战的RNN变体。然而,目前缺乏能够系统测试这些变体的核心库,重新实现多样化的网络架构既耗时又易出错,限制了研究的可复现性与探索效率。为此,我们推出了三个基于Julia与Python的开源库,集中实现了多种循环单元及高层循环架构。torchrecurrent、RecurrentLayers.jl与LuxRecurrentLayers.jl为构建与扩展RNN模型提供了统一框架,并内置了可定制化与实验性功能。所有软件包均采用MIT许可证,并在GitHub上持续维护。