Efficiently capturing the long-range patterns in sequential data sources salient to a given task -- such as classification and generative modeling -- poses a fundamental challenge. Popular approaches in the space tradeoff between the memory burden of brute-force enumeration and comparison, as in transformers, the computational burden of complicated sequential dependencies, as in recurrent neural networks, or the parameter burden of convolutional networks with many or large filters. We instead take inspiration from wavelet-based multiresolution analysis to define a new building block for sequence modeling, which we call a MultiresLayer. The key component of our model is the multiresolution convolution, capturing multiscale trends in the input sequence. Our MultiresConv can be implemented with shared filters across a dilated causal convolution tree. Thus it garners the computational advantages of convolutional networks and the principled theoretical motivation of wavelet decompositions. Our MultiresLayer is straightforward to implement, requires significantly fewer parameters, and maintains at most a $\mathcal{O}(N\log N)$ memory footprint for a length $N$ sequence. Yet, by stacking such layers, our model yields state-of-the-art performance on a number of sequence classification and autoregressive density estimation tasks using CIFAR-10, ListOps, and PTB-XL datasets.
翻译:多分辨率卷积记忆的序列建模
Efficiently capturing the long-range patterns in sequential data sources salient to a given task -- such as classification and generative modeling -- poses a fundamental challenge. Popular approaches in the space tradeoff between the memory burden of brute-force enumeration and comparison, as in transformers, the computational burden of complicated sequential dependencies, as in recurrent neural networks, or the parameter burden of convolutional networks with many or large filters. We instead take inspiration from wavelet-based multiresolution analysis to define a new building block for sequence modeling, which we call a MultiresLayer.
高效地捕捉与给定任务相关的顺序数据源中的长程模式(例如分类和生成建模)是一个根本性的挑战。在这一领域中,流行的方法在存储器负担、计算负担或过多或过大的卷积网络参数负担之间进行平衡处理。我们反而从基于小波多分辨率分析的灵感中提取出一个新的序列建模单元,称之为 MultiresLayer。
The key component of our model is the multiresolution convolution, capturing multiscale trends in the input sequence. Our MultiresConv can be implemented with shared filters across a dilated causal convolution tree. Thus it garners the computational advantages of convolutional networks and the principled theoretical motivation of wavelet decompositions. Our MultiresLayer is straightforward to implement, requires significantly fewer parameters, and maintains at most a $\mathcal{O}(N\log N)$ memory footprint for a length $N$ sequence. Yet, by stacking such layers, our model yields state-of-the-art performance on a number of sequence classification and autoregressive density estimation tasks using CIFAR-10, ListOps, and PTB-XL datasets.
我们模型的关键组成部分是多分辨率卷积,能够捕获输入序列中的多尺度趋势。我们的 MultiresConv 可以在扩张因果卷积树上使用共享的过滤器来实现。因此,它兼具卷积网络的计算优势和小波分解的理论原则动机。我们的 MultiresLayer 实现简单,需要的参数数量明显较少,并且在长度为 N 的序列中最多维护一个 $\mathcal{O}(N\log N)$ 的内存占用。然而,通过叠加这样的层,我们的模型在使用 CIFAR-10、ListOps 和 PTB-XL 数据集进行多个序列分类和自回归密度估计任务方面取得了最先进的性能。