Nowadays new technologies, and especially artificial intelligence, are more and more established in our society. Big data analysis and machine learning, two sub-fields of artificial intelligence, are at the core of many recent breakthroughs in many application fields (e.g., medicine, communication, finance, ...), including some that are strongly related to our day-to-day life (e.g., social networks, computers, smartphones, ...). In machine learning, significant improvements are usually achieved at the price of an increasing computational complexity and thanks to bigger datasets. Currently, cutting-edge models built by the most advanced machine learning algorithms typically became simultaneously very efficient and profitable but also extremely complex. Their complexity is to such an extent that these models are commonly seen as black-boxes providing a prediction or a decision which can not be interpreted or justified. Nevertheless, whether these models are used autonomously or as a simple decision-making support tool, they are already being used in machine learning applications where health and human life are at stake. Therefore, it appears to be an obvious necessity not to blindly believe everything coming out of those models without a detailed understanding of their predictions or decisions. Accordingly, this thesis aims at improving the interpretability of models built by a specific family of machine learning algorithms, the so-called tree-based methods. Several mechanisms have been proposed to interpret these models and we aim along this thesis to improve their understanding, study their properties, and define their limitations.
翻译:当今的新技术,特别是人工智能,在社会上越来越成熟。大数据分析和机器学习这两个人工智能的子领域,是许多应用领域(例如医学、通信、金融.)最近许多突破的核心,其中包括一些与我们日常生活密切相关的技术(例如社交网络、计算机、智能手机,......)。在机器学习中,通常以计算复杂程度不断提高的价格和由于数据组的扩大而实现重大改进。目前,由最先进的机器学习算法所建立的尖端模型通常同时变得非常高效和有利可图,但也非常复杂。这些模型的复杂性是,这些模型通常被视为黑箱,提供预测或决定,不能解释或合理。然而,这些模型是自主使用,还是简单的决策支持工具,已经在机器学习应用中,其中涉及健康和人类生活。因此,显然没有必要盲目地相信这些模型的所有结果都是在没有详细理解其预测和解释的情况下产生的,因此,这些模型的目的就是根据这些预测或机算的精确解释机制来改进。因此,这些模型的目的就是通过这些解释方法的几种解释方法来改进。我们用这些解释,这些方法的目的是要改进这些解释。