We present a preliminary study on an end-to-end variational autoencoder (VAE) for sound morphing. Two VAE variants are compared: VAE with dilation layers (DC-VAE) and VAE only with regular convolutional layers (CC-VAE). We combine the following loss functions: 1) the time-domain mean-squared error for reconstructing the input signal, 2) the Kullback-Leibler divergence to the standard normal distribution in the bottleneck layer, and 3) the classification loss calculated from the bottleneck representation. On a database of spoken digits, we use 1-nearest neighbor classification to show that the sound classes separate in the bottleneck layer. We introduce the Mel-frequency cepstrum coefficient dynamic time warping (MFCC-DTW) deviation as a measure of how well the VAE decoder projects the class center in the latent (bottleneck) layer to the center of the sounds of that class in the audio domain. In terms of MFCC-DTW deviation and 1-NN classification, DC-VAE outperforms CC-VAE. These results for our parametrization and our dataset indicate that DC-VAE is more suitable for sound morphing than CC-VAE, since the DC-VAE decoder better preserves the topology when mapping from the audio domain to the latent space. Examples are given both for morphing spoken digits and drum sounds.
翻译:我们展示了用于音质变形的端到端变异自动读数器(VAE)的初步研究。两种VAE变量比较了:VAE与变相层(DC-VAE),VAE只与普通变相层(CC-VAE)比较。我们将以下损失函数组合在一起:1)重建输入信号的时间-偏差平均偏差,2)重塑输入信号的时间-偏差(VAE),2)Kullback-leeper对瓶头层标准正常分布的偏差,3)从瓶颈代表处计算出的分类损失。在语音数字代表处数据库中,我们使用1个最接近的声频邻居分类来显示瓶状层的音频类别(DC-VAE),我们引入了Mel-频 Cepstruum 系数动态时间扭曲(MFCC-DW)的偏差,以衡量VAE类中心在潜值层(bottnneck)到该类声音中心的中心。在MFCC-D-DW偏离和1NNUER 分类中, 显示我们的上层数据比DC-VAAFA值更适合的DC-VAA值。