The combination of deep neural nets and theory-driven models, which we call deep grey-box modeling, can be inherently interpretable to some extent thanks to the theory backbone. Deep grey-box models are usually learned with a regularized risk minimization to prevent a theory-driven part from being overwritten and ignored by a deep neural net. However, an estimation of the theory-driven part obtained by uncritically optimizing a regularizer can hardly be trustworthy when we are not sure what regularizer is suitable for the given data, which may harm the interpretability. Toward a trustworthy estimation of the theory-driven part, we should analyze regularizers' behavior to compare different candidates and to justify a specific choice. In this paper, we present a framework that enables us to analyze a regularizer's behavior empirically with a slight change in the neural net's architecture and the training objective.
翻译:深神经网和理论驱动模型(我们称之为深灰箱模型)的结合,可以在某种程度上由理论骨干来解释。深灰箱模型通常以正规化的风险最小化方式学习,以防止理论驱动的部分被过度翻写和被深神经网忽视。然而,对理论驱动部分通过不严格优化常规化器获得的估计,当我们不确定什么常规化器适合给定数据,从而可能损害可解释性时,很难令人信服。为了对理论驱动部分进行可信的估计,我们应该分析规范化者的行为,比较不同的候选人,并解释具体选择的理由。在本文中,我们提出了一个框架,使我们能够通过对神经网结构和培训目标稍作改变,对常规化者的行为进行实证分析。