We present a novel hybrid sound propagation algorithm for interactive applications. Our approach is designed for general dynamic scenes and uses a neural network-based learned scattered field representation along with ray tracing to generate specular, diffuse, diffraction, and occlusion effects efficiently. To handle general objects, we exploit properties of the acoustic scattering field and use geometric deep learning to approximate the field using spherical harmonics. We use a large dataset for training, and compare its accuracy with the ground truth generated using an accurate wave-based solver. The additional overhead of computing the learned scattered field at runtime is small and we highlight the interactive performance by generating plausible sound effects in dynamic scenes with diffraction and occlusion effects. We demonstrate the perceptual benefits of our approach based on an audio-visual user study.
翻译:我们为互动应用程序提供了一个新型混合声音传播算法。 我们的方法是为一般动态场景设计的,并使用基于神经网络的已知分散场面以及射线追踪来高效生成光学、扩散、分解和隔绝效应。 为了处理普通物体,我们利用声学散射场的特性,并利用球形口音进行几何深识学,以近似现场。 我们使用庞大的数据集进行培训,并用精确的波形解答器来比较其准确性与地面真相。 在运行时计算已知分散场的额外间接费用很小,我们强调互动性能,在动态场景中产生貌似声音效果,产生偏差和隔绝效应。我们展示了我们基于视听用户研究的方法在概念上的好处。