Explainability of neural network prediction is essential to understand feature importance and gain interpretable insight into neural network performance. In this work, model explanations are fed back to the feed-forward training to help the model generalize better. To this extent, a custom weighted loss where the weights are generated by considering the Euclidean distances between true LIME (Local Interpretable Model-Agnostic Explanations) explanations and model-predicted LIME explanations is proposed. Also, in practical training scenarios, developing a solution that can help the model learn sequentially without losing information on previous data distribution is imperative due to the unavailability of all the training data at once. Thus, the framework known as XAI-Increment incorporates the custom weighted loss developed with elastic weight consolidation (EWC), to maintain performance in sequential testing sets. Finally, the training procedure involving the custom weighted loss shows around 1% accuracy improvement compared to the traditional loss based training for the keyword spotting task on the Google Speech Commands dataset and also shows low loss of information when coupled with EWC in the incremental learning setup.
翻译:神经网络预测的可解释性对于理解特征重要性和获得对神经网络性能的可解释的洞察力至关重要。在这项工作中,示范解释被反馈到进化前培训中,帮助模型更全面地概括模型。在这方面,考虑到真正的LIME(当地解释模型-不可想象的解释)解释和模型预测的LIME解释之间的ECLE距离而产生加权效应的定制加权损失。此外,在实际培训情景中,开发一种解决方案,帮助模型在不丢失先前数据分布信息的情况下按顺序学习,由于一次性得不到所有培训数据,这种解决方案十分必要。因此,称为XAI-Incenterment 的框架纳入了与弹性重量合并(EWC)开发的定制加权损失,以保持连续测试组的性能。最后,与基于谷歌语音指令数据集关键识别任务培训的传统损失相比,涉及定制加权损失的精确度提高了1%左右,并且与EWC一起在递增学习设置中的信息损失较低。