Despite the considerable success of neural networks in security settings such as malware detection, such models have proved vulnerable to evasion attacks, in which attackers make slight changes to inputs (e.g., malware) to bypass detection. We propose a novel approach, \emph{Fourier stabilization}, for designing evasion-robust neural networks with binary inputs. This approach, which is complementary to other forms of defense, replaces the weights of individual neurons with robust analogs derived using Fourier analytic tools. The choice of which neurons to stabilize in a neural network is then a combinatorial optimization problem, and we propose several methods for approximately solving it. We provide a formal bound on the per-neuron drop in accuracy due to Fourier stabilization, and experimentally demonstrate the effectiveness of the proposed approach in boosting robustness of neural networks in several detection settings. Moreover, we show that our approach effectively composes with adversarial training.
翻译:尽管在恶意软件检测等安全环境下神经网络取得了相当大的成功,但这类模型证明很容易受到规避袭击,其中攻击者对绕过检测的投入(如恶意软件)略作改动。我们提出了一种新颖的方法,即\emph{Fourier 稳定 },用二进制投入设计规避机器人神经网络。这一方法是对其他形式的防御的补充,用Fourier 分析工具来替代个体神经元的重量。选择神经元稳定在神经网络中是一个组合优化问题,我们提出几种方法来大致解决这个问题。我们提供了四级稳定后每中风准确性下降的正式约束,并实验性地展示了拟议方法在几个探测环境中增强神经网络稳健性的有效性。此外,我们展示了我们的方法与对抗性培训的有效结合。