Deep Learning models have become potential candidates for auditory neuroscience research, thanks to their recent successes on a variety of auditory tasks. Yet, these models often lack interpretability to fully understand the exact computations that have been performed. Here, we proposed a parametrized neural network layer, that computes specific spectro-temporal modulations based on Gabor kernels (Learnable STRFs) and that is fully interpretable. We evaluated predictive capabilities of this layer on Speech Activity Detection, Speaker Verification, Urban Sound Classification and Zebra Finch Call Type Classification. We found out that models based on Learnable STRFs are on par for all tasks with different toplines, and obtain the best performance for Speech Activity Detection. As this layer is fully interpretable, we used quantitative measures to describe the distribution of the learned spectro-temporal modulations. The filters adapted to each task and focused mostly on low temporal and spectral modulations. The analyses show that the filters learned on human speech have similar spectro-temporal parameters as the ones measured directly in the human auditory cortex. Finally, we observed that the tasks organized in a meaningful way: the human vocalizations tasks closer to each other and bird vocalizations far away from human vocalizations and urban sounds tasks.
翻译:深学习模型因其在各种听力任务方面最近的成功而成为了听觉神经科学研究的潜在候选体。然而,这些模型往往缺乏解释性,无法充分理解所完成的精确计算。在这里,我们提议了一个准美化的神经网络层,根据Gabor 内核(Learnable Storfors)计算出具体的光谱-时空调调调制,这是完全可以解释的。我们评估了这一层的预测能力,涉及语音活动探测、议长核查、城市健全分类和Zebra Finch调用类型分类。我们发现,基于可学习的STRF的模型与所有具有不同顶线的任务完全相同,并获得了语音活动探测的最佳性能。由于这个层完全可以解释,我们使用了定量措施来描述所学的光谱-时空调制调制的分布。根据每项任务调整的过滤器,主要侧重于低时光调和光谱调调制。分析显示,关于人类演讲的过滤器的光谱-时空调参数与直接测量到人类语音任务的最远处,我们观察了每个语音任务。