The concept of trustworthy AI has gained widespread attention lately. One of the aspects relevant to trustworthy AI is robustness of ML models. In this study, we show how to probabilistically quantify robustness against naturally occurring distortions of input data for tree-based classifiers under the assumption that the natural distortions can be described by multivariate probability distributions that can be transformed to multivariate normal distributions. The idea is to extract the decision rules of a trained tree-based classifier, separate the feature space into non-overlapping regions and determine the probability that a data sample with distortion returns its predicted label. The approach is based on the recently introduced measure of real-world-robustness, which works for all black box classifiers, but is only an approximation and only works if the input dimension is not too high, whereas our proposed method gives an exact measure.
翻译:值得信赖的AI概念最近得到了广泛的关注。 值得信赖的AI概念的一个相关方面是 ML 模型的稳健性。 在这项研究中,我们展示了如何以概率量化稳健性,防止自然发生的树类分类数据输入数据扭曲,假设自然扭曲可以通过多变概率分布来描述,这种分布可以转换成多变正常分布。 其想法是提取训练有素的树类分类师的决策规则,将特征空间分离成非重叠区域,并确定带有扭曲性的数据样本返回其预测标签的可能性。 这种方法基于最近引入的实生世界-紫色度测量法,该测量法适用于所有黑盒分类师,但只是近似值,只有在输入尺寸不高的情况下才起作用,而我们提出的方法则给出了精确的尺度。