Distributed machine learning paradigms, such as federated learning, have been recently adopted in many privacy-critical applications for speech analysis. However, such frameworks are vulnerable to privacy leakage attacks from shared gradients. Despite extensive efforts in the image domain, the exploration of speech privacy leakage from gradients is quite limited. In this paper, we explore methods for recovering private speech/speaker information from the shared gradients in distributed learning settings. We conduct experiments on a keyword spotting model with two different types of speech features to quantify the amount of leaked information by measuring the similarity between the original and recovered speech signals. We further demonstrate the feasibility of inferring various levels of side-channel information, including speech content and speaker identity, under the distributed learning framework without accessing the user's data.
翻译:最近,许多对隐私至关重要的语音分析应用中都采用了分布式机器学习模式,如联合学习,但这类框架很容易受到来自共享梯度的隐私泄漏攻击。尽管在图像领域作出了大量努力,但探索从梯度中语音隐私泄漏的情况非常有限。在本文中,我们探讨了从分布式学习环境中的共享梯度中恢复私人语音/语音信息的方法。我们用一种关键词识别模式进行实验,该模式有两种不同的语音特征,通过测量原始和已恢复的语音信号的相似性来量化泄漏信息的数量。我们进一步展示了在分布式学习框架内在没有访问用户数据的情况下推断不同层次的侧道信息,包括语音内容和发言者身份的可行性。