Software Log anomaly event detection with masked event prediction has various technical approaches with countless configurations and parameters. Our objective is to provide a baseline of settings for similar studies in the future. The models we use are the N-Gram model, which is a classic approach in the field of natural language processing (NLP), and two deep learning (DL) models long short-term memory (LSTM) and convolutional neural network (CNN). For datasets we used four datasets Profilence, BlueGene/L (BGL), Hadoop Distributed File System (HDFS) and Hadoop. Other settings are the size of the sliding window which determines how many surrounding events we are using to predict a given event, mask position (the position within the window we are predicting), the usage of only unique sequences, and the portion of data that is used for training. The results show clear indications of settings that can be generalized across datasets. The performance of the DL models does not deteriorate as the window size increases while the N-Gram model shows worse performance with large window sizes on the BGL and Profilence datasets. Despite the popularity of Next Event Prediction, the results show that in this context it is better not to predict events at the edges of the subsequence, i.e., first or last event, with the best result coming from predicting the fourth event when the window size is five. Regarding the amount of data used for training, the results show differences across datasets and models. For example, the N-Gram model appears to be more sensitive toward the lack of data than the DL models. Overall, for similar experimental setups we suggest the following general baseline: Window size 10, mask position second to last, do not filter out non-unique sequences, and use a half of the total data for training.
翻译:以隐形事件预测来测算软件的异常事件有各种技术方法。 我们的目标是为未来的类似研究提供一个设置基准。 我们使用的模式是 N- Gram 模型, 这是自然语言处理( NLP) 领域的经典方法, 以及两个深度学习( DL) 模型, 长期短期内存( LSTM) 和 convolual 神经网络( CNN ) 。 对于数据集, 我们使用了四个数据集 Profile、 BlueGene/ L (BGL)、 Hadoop 分配文件系统 (HDFS) 和 Hadoop 。 其他的设置是滑动窗口的大小, 用于预测一个特定事件、 掩码位置( 我们预测的窗口内的位置) 、 仅使用独有序列的模型以及用于培训的部分。 结果表明, 可在数据集的最后一个设置中, 当窗口大小增加时, DLeL 格式模式的性能不会恶化, 而 N- gram 模式显示在下一个窗口的大小, 在 BGL 和 Gread Riqueal 中, 数据显示 的缩 的底 显示 数据 显示 的大小。