The use of machine learning (ML) models in decision-making contexts, particularly those used in high-stakes decision-making, are fraught with issue and peril since a person - not a machine - must ultimately be held accountable for the consequences of the decisions made using such systems. Machine learning explainability (MLX) promises to provide decision-makers with prediction-specific rationale, assuring them that the model-elicited predictions are made for the right reasons and are thus reliable. Few works explicitly consider this key human-in-the-loop (HITL) component, however. In this work we propose HEX, a human-in-the-loop deep reinforcement learning approach to MLX. HEX incorporates 0-distrust projection to synthesize decider specific explanation-providing policies from any arbitrary classification model. HEX is also constructed to operate in limited or reduced training data scenarios, such as those employing federated learning. Our formulation explicitly considers the decision boundary of the ML model in question, rather than the underlying training data, which is a shortcoming of many model-agnostic MLX methods. Our proposed methods thus synthesize HITL MLX policies that explicitly capture the decision boundary of the model in question for use in limited data scenarios.
翻译:机械学习解释(MLX)承诺向决策者提供具体预测的理由,向他们保证模型允许的预测是出于正确的理由,因此是可靠的。很少有工作明确考虑到这种关键的“人与圈(HITL)”(HITL)部分的关键“人与圈(HITL)”部分。在这项工作中,我们提出“HEX”,这是MLX的“人与圈(HEX)的深层强化学习方法”。HEX包含0-不信任预测,以综合任何任意分类模式中的决定者具体解释提供的政策。HEX还设计在有限或减少的培训数据假设中运作,例如采用“Federated”学习。我们的提法明确考虑了有关的“模型”模式的决策界限,而不是基本的培训数据,这是许多“模型”MLX方法的一个短处。我们提出的方法如此综合了“HITLLL MLX”模型中的有限数据模型,从而明确测量了决定的边界问题。