Machine learning systems often do not share the same inductive biases as humans and, as a result, extrapolate or generalize in ways that are inconsistent with our expectations. The trade-off between exemplar- and rule-based generalization has been studied extensively in cognitive psychology; in this work, we present a protocol inspired by these experimental approaches to probe the inductive biases that control this tradeoff in category-learning systems. We isolate two such inductive biases: feature-level bias (differences in which features are more readily learned) and exemplar or rule bias (differences in how these learned features are used for generalization). We find that standard neural network models are feature-biased and exemplar-based, and discuss the implications of these findings for machine learning research on systematic generalization, fairness, and data augmentation.
翻译:机器学习系统往往与人类不具有相同的感官偏见,因此,以不符合我们期望的方式外推或概括。在认知心理学中广泛研究了基于理论的概括和基于规则的概括之间的权衡;在这项工作中,我们提出了一个由这些实验方法启发的协议,以探究控制分类学习系统中这种权衡的暗示偏见。我们分离了两种这样的感官偏见:特征偏差(特征更易于学习的不同)和实例或规则偏差(这些所学特征如何用于概括的不同)。我们发现标准神经网络模型具有特征偏向性和基于实例的平衡,并讨论了这些发现对系统普及、公平和数据增强的机器学习研究的影响。