Context: In recent years, leveraging machine learning (ML) techniques has become one of the main solutions to tackle many software engineering (SE) tasks, in research studies (ML4SE). This has been achieved by utilizing state-of-the-art models that tend to be more complex and black-box, which is led to less explainable solutions that reduce trust and uptake of ML4SE solutions by professionals in the industry. Objective: One potential remedy is to offer explainable AI (XAI) methods to provide the missing explainability. In this paper, we aim to explore to what extent XAI has been studied in the SE community (XAI4SE) and provide a comprehensive view of the current state-of-the-art as well as challenge and roadmap for future work. Method: We conduct a systematic literature review on 24 (out of 869 primary studies that were selected by keyword search) most relevant published studies in XAI4SE. We have three research questions that were answered by meta-analysis of the collected data per paper. Results: Our study reveals that among the identified studies, software maintenance (\%68) and particularly defect prediction has the highest share on the SE stages and tasks being studied. Additionally, we found that XAI methods were mainly applied to classic ML models rather than more complex models. We also noticed a clear lack of standard evaluation metrics for XAI methods in the literature which has caused confusion among researchers and a lack of benchmarks for comparisons. Conclusions: XAI has been identified as a helpful tool by most studies, which we cover in the systematic review. However, XAI4SE is a relatively new domain with a lot of untouched potentials, including the SE tasks to help with, the ML4SE methods to explain, and the types of explanations to offer. This study encourages the researchers to work on the identified challenges and roadmap reported in the paper.
翻译:近些年来,利用机器学习(ML)技术已成为解决许多软件工程(SE)任务的主要解决办法之一,这是研究研究(ML4SE)中(ML4SE)中许多软件工程(SE)任务的主要解决办法之一。这是通过使用最先进的模型实现的,这些模型往往更为复杂和黑箱,导致较不易解释的解决方案,降低行业专业人员对ML4SE解决方案的信任和吸收程度。目标:一个潜在的补救措施是提供可解释的 AI (XAI) 方法,以提供缺失的可解释性解释性解释。在本文中,我们旨在探索SEAI(XAI4SE)在SE社区(XAI4SESE)中研究过多少程度,并提供了对目前SEA最新状态以及未来工作的挑战和路线图的系统审查。方法:我们对24个(869项初级研究中,通过关键搜索选定了869项基本研究)进行了系统文献审查,对XAI行业专业人员对所收集的数据进行了最相关的研究。我们通过对所收集的数据进行元分析,提供了三个研究的答案解答。结果:我们的研究显示,我们发现,在所发现的潜在基准中发现,在所发现的研究中,在所发现,软件中,软件维护(68中,特别是测测测测测读中,在SIT研究中,在SAI研究中,在SIM研究中,在SIM研究中,在SIM研究中,在SIS研究中,在SIS研究中采用了最高比例中,在SIS研究中,在SIS研究中,在SIS研究中,在SIS研究中,在SDR研究中,在SDR研究中,以及任务中,在SER中,在SD中,在SD中,在SEA中,在SD中,以及任务中,在SD中,以及任务中,在S的计算中,在SD中,在SD中,在研究中,在研究中,在研究中,在SL研究中也缺少中也缺少中,在SL研究中,我们报告了最高分中,也缺乏中,在研究中,我们报告了最高比例中,在研究中,在研究中,在研究中,以及任务中,在研究中,在研究中,以及缺乏中