How can we explain the predictions of a machine learning model? When the data is structured as a multivariate time series, this question induces additional difficulties such as the necessity for the explanation to embody the time dependency and the large number of inputs. To address these challenges, we propose dynamic masks (Dynamask). This method produces instance-wise importance scores for each feature at each time step by fitting a perturbation mask to the input sequence. In order to incorporate the time dependency of the data, Dynamask studies the effects of dynamic perturbation operators. In order to tackle the large number of inputs, we propose a scheme to make the feature selection parsimonious (to select no more feature than necessary) and legible (a notion that we detail by making a parallel with information theory). With synthetic and real-world data, we demonstrate that the dynamic underpinning of Dynamask, together with its parsimony, offer a neat improvement in the identification of feature importance over time. The modularity of Dynamask makes it ideal as a plug-in to increase the transparency of a wide range of machine learning models in areas such as medicine and finance, where time series are abundant.
翻译:我们如何解释机器学习模型的预测?当数据结构是一个多变的时间序列时,这一问题又引起更多的困难,例如有必要解释如何体现时间依赖和大量投入。为了应对这些挑战,我们提议使用动态面具(Dynamask)。这种方法通过将扰动面具与输入序列相配,为每个特征的每个阶段产生与实例相容的重要性评分。为了纳入数据的时间依赖,Dynamask研究动态扰动操作器的影响。为了应对大量投入,我们提议了一个方案,使特征选择选项具有针对性(选择不需要的特性)和可理解性(我们通过与信息理论平行进行详细描述的理念 ) 。 在合成数据和现实世界数据中,我们证明Dynamask的动态基础,连同其微调,能够很好地改善对时间特征重要性的识别。Dynamask的模块化使得它成为理想的插座,可以增加医学和金融等领域一系列机器学习模型的透明度,其中的时间是丰富的。