Neural network stealing attacks have posed grave threats to neural network model deployment. Such attacks can be launched by extracting neural architecture information, such as layer sequence and dimension parameters, through leaky side-channels. To mitigate such attacks, we propose NeurObfuscator, a full-stack obfuscation tool to obfuscate the neural network architecture while preserving its functionality with very limited performance overhead. At the heart of this tool is a set of obfuscating knobs, including layer branching, layer widening, selective fusion and schedule pruning, that increase the number of operators, reduce/increase the latency, and number of cache and DRAM accesses. A genetic algorithm-based approach is adopted to orchestrate the combination of obfuscating knobs to achieve the best obfuscating effect on the layer sequence and dimension parameters so that the architecture information cannot be successfully extracted. Results on sequence obfuscation show that the proposed tool obfuscates a ResNet-18 ImageNet model to a totally different architecture (with 44 layer difference) without affecting its functionality with only 2% overall latency overhead. For dimension obfuscation, we demonstrate that an example convolution layer with 64 input and 128 output channels can be obfuscated to generate a layer with 207 input and 93 output channels with only a 2% latency overhead.
翻译:神经网络盗窃袭击对神经网络模型部署构成了严重威胁。这种袭击可以通过提取神经结构信息,如层序列和维度参数等神经结构信息,通过漏泄的侧通道,通过层序列和维度参数等提取神经结构信息。为了减轻这些袭击,我们提议 NeurObfuscator, 这是一种全斯塔式的模糊模糊工具, 用来混淆神经网络结构结构, 同时用非常有限的性能管理来维护其功能, 其核心是一套令人困惑的软体, 包括层分支、 层扩大、 有选择的聚合和时间表运行, 从而增加操作者的数量, 减少/ 增加延度, 以及缓存和 DRAM 访问的次数。 为了减轻这些袭击, 我们建议采用基于遗传算法的方法, 将固化 knobs 组合起来, 以在层序列和尺寸参数上取得最佳的模糊效果, 从而无法成功提取建筑信息。 解析的序列结果显示, 拟议的工具只能将ResNet-18 图像网络模型解析成一个完全不同的结构(有44层的粘度差异), 其功能只能通过2 的平层平面结构演示演示演示演示,, 只能显示我们的平层平层输出, 。