Machine learning models and algorithms are used in a number of systems that affect our daily life. Thus, in some settings, methods that are easy to explain or interpret may be highly desirable. The price of explainability can be thought of as the loss in terms of quality that is unavoidable if we restrict these systems to use explainable methods. We study the price of explainability, under a theoretical perspective, for clustering tasks. We provide upper and lower bounds on this price as well as efficient algorithms to build explainable clustering for the $k$-means, $k$-medians, $k$-center and the maximum-spacing problems in a natural model in which explainability is achieved via decision trees.
翻译:许多影响我们日常生活的系统都使用机器学习模式和算法,因此,在某些环境中,容易解释或解释的方法可能非常可取,如果限制这些系统使用可解释的方法,解释的代价可视为质量损失,而这种损失不可避免。我们从理论角度研究可解释性的代价,以便进行集群任务。我们对这一价格提供上下限,并提供高效的算法,以建立可解释的组合,包括美元-人民币、美元-中间货币、美元-中间货币和最高间隔问题,在自然模型中,通过决策树实现解释性。