Neural speech codecs aim to compress input signals into minimal bits while maintaining content quality in a low-latency manner. However, existing codecs often trade computational complexity for reconstruction performance. These codecs primarily use convolutional blocks for feature transformation layers, which are not inherently suited for capturing the local redundancies in speech signals. To compensate, they require either adversarial discriminators or a large number of model parameters to enhance audio quality. In response to these challenges, we introduce the Efficient Speech Codec (ESC), a lightweight, parameter-efficient speech codec based on a cross-scale residual vector quantization scheme and transformers. Our model employs mirrored hierarchical window transformer blocks and performs step-wise decoding from coarse-to-fine feature representations. To enhance bitrate efficiency, we propose a novel combination of vector quantization techniques along with a pre-training paradigm. Extensive experiments demonstrate that ESC can achieve high-fidelity speech reconstruction with significantly lower complexity, making it a promising alternative to existing convolutional audio codecs.
翻译:暂无翻译