Translating machine learning algorithms into clinical applications requires addressing challenges related to interpretability, such as accounting for the effect of confounding variables (or metadata). Confounding variables affect the relationship between input training data and target outputs. When we train a model on such data, confounding variables will bias the distribution of the learned features. A recent promising solution, MetaData Normalization (MDN), estimates the linear relationship between the metadata and each feature based on a non-trainable closed-form solution. However, this estimation is confined by the sample size of a mini-batch and thereby may cause the approach to be unstable during training. In this paper, we extend the MDN method by applying a Penalty approach (referred to as PDMN). We cast the problem into a bi-level nested optimization problem. We then approximate this optimization problem using a penalty method so that the linear parameters within the MDN layer are trainable and learned on all samples. This enables PMDN to be plugged into any architectures, even those unfit to run batch-level operations, such as transformers and recurrent models. We show improvement in model accuracy and greater independence from confounders using PMDN over MDN in a synthetic experiment and a multi-label, multi-site dataset of magnetic resonance images (MRIs).
翻译:将机器学习算法转换为临床应用需要解决与可解释性有关的挑战,例如计算混杂变量(或元数据)的影响。混杂变量影响投入培训数据和目标产出之间的关系。当我们培训一个数据模型时,混杂变量将偏向于所学特征的分布。最近的一个有希望的解决办法Metadata 正常化(MDN),根据非可控制封闭式封闭式解决方案估算元数据和每个特征之间的线性关系。然而,这一估计受小型批量样本大小的限制,从而可能导致在培训期间采用的方法不稳定。在本文中,我们通过应用惩罚方法(称为PDMN)来扩展MDN方法。我们把问题推入双层嵌套式优化问题。然后我们用惩罚方法来弥补这一优化问题,以便MDN层内的线性参数可以被培训和学习。这使得PMDN能够连接到任何结构中,甚至那些不适于进行批量级操作的架构,例如变压器和经常模型。我们展示了模型的精确性和更大独立性,我们展示了模型的精度,从合成图像的合成NMDMDR的多级试验中,我们展示了比合成图像的模型的模型的模型更独立性。