The goal of model distillation is to faithfully transfer teacher model knowledge to a model which is faster, more generalizable, more interpretable, or possesses other desirable characteristics. Human-readability is an important and desirable standard for machine-learned model interpretability. Readable models are transparent and can be reviewed, manipulated, and deployed like traditional source code. As a result, such models can be improved outside the context of machine learning and manually edited if desired. Given that directly training such models is difficult, we propose to train interpretable models using conventional methods, and then distill them into concise, human-readable code. The proposed distillation methodology approximates a model's univariate numerical functions with piecewise-linear curves in a localized manner. The resulting curve model representations are accurate, concise, human-readable, and well-regularized by construction. We describe a piecewise-linear curve-fitting algorithm that produces high-quality results efficiently and reliably across a broad range of use cases. We demonstrate the effectiveness of the overall distillation technique and our curve-fitting algorithm using three publicly available datasets COMPAS, FICO, and MSLR-WEB30K.
翻译:模型蒸馏的目的是忠实地将教师模型知识转让给一种更快、更普遍、更可解释或具有其他可取特征的模型; 人类可读性是机器学习模型解释性的一个重要和可取的标准; 可读性模型透明,可以像传统源代码一样加以审查、操纵和部署; 因此,可以在机器学习之外改进这些模型,如果需要的话,可以人工编辑; 鉴于直接培训这类模型是困难的,我们提议用传统方法来培训可解释的模型,然后用简单、易读的代码加以提炼; 提议的蒸馏方法以局部方式将模型的单象形数字功能与片线曲线曲线相近; 由此形成的曲线模型表示准确、简洁、可读、易于操作,并按结构进行正规化。 我们描述了在各种使用案例中高效和可靠地产生高质量结果的细线调算法。 我们展示了总体蒸馏技术的有效性,以及我们使用三种公开数据集COMAS、FICO和MSLR-WE30K的曲线校准算法。