This paper compares two deep reinforcement learning approaches for cyber security in software defined networking. Neural Episodic Control to Deep Q-Network has been implemented and compared with that of Double Deep Q-Networks. The two algorithms are implemented in a format similar to that of a zero-sum game. A two-tailed T-test analysis is done on the two game results containing the amount of turns taken for the defender to win. Another comparison is done on the game scores of the agents in the respective games. The analysis is done to determine which algorithm is the best in game performer and whether there is a significant difference between them, demonstrating if one would have greater preference over the other. It was found that there is no significant statistical difference between the two approaches.
翻译:本文比较了软件定义网络的网络安全的两个深度强化学习方法。 对深QNetwork 的神经Episodic Control已经实施, 并与双深QNetwork 比较。 这两种算法都以类似于零和游戏的格式实施。 对包含辩护人获胜所需转折次数的两种游戏结果进行了双尾T测试分析。 对各自游戏中代理商的游戏分数进行了另一个比较。 分析的目的是确定哪种算法在游戏表演中是最佳的, 以及两者之间是否有重大差异, 表明一种算法是否会比另一种更偏好。 发现两种方法之间没有重大的统计差异 。