Deep-learning models such as Convolutional Neural Networks (CNN) and Long Short-Term Memory (LSTM) have been successfully used for process-mining tasks. They have achieved better performance for different predictive tasks than traditional approaches. We extend the existing body of research by testing four different variants of Graph Neural Networks (GNN) and a fully connected Multi-layer Perceptron (MLP) with dropout for the tasks of predicting the nature and timestamp of the next process activity. In contrast to existing studies, we evaluate our models' performance at different stages of a process, determined by quartiles of the number of events and normalized quarters of the case duration. This provides new insights into the performance of a prediction model, as they behave differently at different stages of a business-process. Interestingly, our experiments show that the simple MLP often outperforms more sophisticated deep-learning models in both prediction tasks. We argue that care needs to be taken when applying automated process-prediction techniques at different stages of a process. We further argue that researchers should reflect their results with strong baselines methods like simple MLPs.
翻译:革命神经网络(CNN)和长期短期内存(LSTM)等深层学习模型被成功地用于处理采矿任务,这些模型比传统方法在不同的预测任务中取得了更好的业绩。我们通过测试四个不同的图表神经网络(GNN)和完全连接的多层感应器(MLP)的四种不同的变体和完全连接的多层感应器(MLP)来测试预测下一个过程活动的性质和时间戳。与现有的研究不同,我们评估了我们的模型在一个由事件数量四分法和案件周期的正常间隔决定的不同过程各阶段的绩效。这为预测模型的性能提供了新的洞见,因为它们在商业过程的不同阶段的表现不同。有意思的是,我们的实验表明,简单的MLP往往在两个预测任务中都比更精密的深学习模型要强。我们主张,在程序的不同阶段应用自动过程定位技术时需要谨慎。我们进一步认为,研究人员应该用简单的MLPs等强有力的基线方法反映其结果。