We present a method of explainable artificial intelligence (XAI), "What I Know (WIK)", to provide additional information to verify the reliability of a deep learning model by showing an example of an instance in a training dataset that is similar to the input data to be inferred and demonstrate it in a remote sensing image classification task. One of the expected roles of XAI methods is verifying whether inferences of a trained machine learning model are valid for an application, and it is an important factor that what datasets are used for training the model as well as the model architecture. Our data-centric approach can help determine whether the training dataset is sufficient for each inference by checking the selected example data. If the selected example looks similar to the input data, we can confirm that the model was not trained on a dataset with a feature distribution far from the feature of the input data. With this method, the criteria for selecting an example are not merely data similarity with the input data but also data similarity in the context of the model task. Using a remote sensing image dataset from the Sentinel-2 satellite, the concept was successfully demonstrated with reasonably selected examples. This method can be applied to various machine-learning tasks, including classification and regression.
翻译:我们提出了一个可解释的人工智能(XAI)方法,即“我所知道的(WIK)”,以提供补充信息,通过在类似于要推断的输入数据并在遥感图像分类任务中加以展示的培训数据集中展示一个实例,来核实深学习模型的可靠性。XAI方法的预期作用之一是核查经过培训的机器学习模型的推论是否对应用程序有效,这是一个重要因素,用于培训模型和模型结构的数据集是哪些。我们以数据为中心的方法可以通过检查选定的示例数据来帮助确定培训数据集是否足以计算出每一项推断。如果选定的示例与输入数据相似,我们可以确认该模型没有在与输入数据特征分布相距遥远的数据集上接受培训。使用这种方法,选择一个示例的标准不仅仅是与输入数据相似的数据,而且还是模型任务中的数据相似性。使用来自Sentinel-2卫星的遥感图像数据集,这个概念通过合理选择的示例得到成功演示。这个方法可以适用于各种回归任务,包括模型。