Federated learning has emerged as a prominent privacy-preserving technique for leveraging large-scale distributed datasets by sharing gradients instead of raw data. However, recent studies indicate that private training data can still be exposed through gradient inversion attacks. While earlier analytical methods have demonstrated success in reconstructing input data from fully connected layers, their effectiveness significantly diminishes when applied to convolutional layers, high-dimensional inputs, and scenarios involving multiple training examples. This paper extends our previous work \cite{eltaras2024r} and proposes three advanced algorithms to broaden the applicability of gradient inversion attacks. The first algorithm presents a novel data leakage method that efficiently exploits convolutional layer gradients, demonstrating that even with non-fully invertible activation functions, such as ReLU, training samples can be analytically reconstructed directly from gradients without the need to reconstruct intermediate layer outputs. Building on this foundation, the second algorithm extends this analytical approach to support high-dimensional input data, substantially enhancing its utility across complex real-world datasets. The third algorithm introduces an innovative analytical method for reconstructing mini-batches, addressing a critical gap in current research that predominantly focuses on reconstructing only a single training example. Unlike previous studies that focused mainly on the weight constraints of convolutional layers, our approach emphasizes the pivotal role of gradient constraints, revealing that successful attacks can be executed with fewer than 5\% of the constraints previously deemed necessary in certain layers.
翻译:联邦学习已成为一种重要的隐私保护技术,它通过共享梯度而非原始数据来利用大规模分布式数据集。然而,近期研究表明,私有训练数据仍可能通过梯度反演攻击被泄露。尽管早期的解析方法已成功从全连接层重建输入数据,但当应用于卷积层、高维输入及涉及多个训练样本的场景时,其有效性显著下降。本文扩展了我们先前的工作\cite{eltaras2024r},提出了三种先进算法以拓宽梯度反演攻击的适用范围。第一种算法提出了一种新颖的数据泄露方法,能高效利用卷积层梯度,证明即使对于非完全可逆的激活函数(如ReLU),训练样本也可直接从梯度解析重建,无需重建中间层输出。在此基础上,第二种算法将此解析方法扩展至支持高维输入数据,显著提升了其在复杂现实数据集中的实用性。第三种算法引入了一种创新的解析方法用于重建小批量数据,解决了当前研究主要集中于重建单个训练样本的关键空白。与先前主要关注卷积层权重约束的研究不同,我们的方法强调梯度约束的关键作用,揭示在某些层中,成功攻击所需的约束量可少于先前认为必要量的5%。