Out-of-distribution detection is one of the most critical issue in the deployment of machine learning. The data analyst must assure that data in operation should be compliant with the training phase as well as understand if the environment has changed in a way that autonomous decisions would not be safe anymore. The method of the paper is based on eXplainable Artificial Intelligence (XAI); it takes into account different metrics to identify any resemblance between in-distribution and out of, as seen by the XAI model. The approach is non-parametric and distributional assumption free. The validation over complex scenarios (predictive maintenance, vehicle platooning, covert channels in cybersecurity) corroborates both precision in detection and evaluation of training-operation conditions proximity. Results are available via open source and open data at the following link: https://github.com/giacomo97cnr/Rule-based-ODD.
翻译:摘要:外部分布检测是机器学习部署过程中最为关键的问题之一。数据分析师必须确保操作中的数据与训练阶段相符,并理解环境是否发生了变化,以便能够保证自主决策的安全性。本文方法基于可解释的人工智能(XAI),考虑了不同的指标来识别内部和外部分布之间的任何相似之处,通过XAI模型进行观察。该方法是非参数化的,不需要分布假定。针对复杂情景(预测性维护、车队编组、网络安全的隐蔽信道)的验证证实了检测精度和训练操作条件接近度的同时。结果可通过以下链接进行开源和开放数据获取:https://github.com/giacomo97cnr/Rule-based-ODD。