Semantic networks provide a useful tool to understand how related concepts are retrieved from memory. However, most current network approaches use pairwise links to represent memory recall patterns. Pairwise connections neglect higher-order associations, i.e. relationships between more than two concepts at a time. These higher-order interactions might covariate with (and thus contain information about) how similar concepts are along psycholinguistic dimensions like arousal, valence, familiarity, gender and others. We overcome these limits by introducing feature-rich cognitive hypergraphs as quantitative models of human memory where: (i) concepts recalled together can all engage in hyperlinks involving also more than two concepts at once (cognitive hypergraph aspect), and (ii) each concept is endowed with a vector of psycholinguistic features (feature-rich aspect). We build hypergraphs from word association data and use evaluation methods from machine learning features to predict concept concreteness. Since concepts with similar concreteness tend to cluster together in human memory, we expect to be able to leverage this structure. Using word association data from the Small World of Words dataset, we compared a pairwise network and a hypergraph with N=3586 concepts/nodes. Interpretable artificial intelligence models trained on (1) psycholinguistic features only, (2) pairwise-based feature aggregations, and on (3) hypergraph-based aggregations show significant differences between pairwise and hypergraph links. Specifically, our results show that higher-order and feature-rich hypergraph models contain richer information than pairwise networks leading to improved prediction of word concreteness. The relation with previous studies about conceptual clustering and compartmentalisation in associative knowledge and human memory are discussed.
翻译:语义网络提供了一个有用的工具,用于了解如何从记忆中检索相关概念。然而,大多数当前的网络方法使用成对的链接来表示记忆检索模式。成对连接忽略了高阶关联,即一次涉及两个以上的概念之间的关系。这些高阶交互可能与类似于唤起、价值、熟悉度、性别等心理语言学维度沿着的概念相似性共变,并蕴含相关信息。通过引入功能丰富的认知超图作为人类记忆的定量模型,我们克服了这些限制,其中: (i) 一起召回的概念可以同时涉及也涉及两个以上的概念的超链接 (认知超图方面),以及 (ii) 每个概念都具有心理语言学特征向量 (功能丰富方面)。我们从单词联想数据构建超图,并使用机器学习特征的评估方法来预测概念具体性。由于具有类似具体性的概念倾向于在人类记忆中聚集在一起,我们期望能够利用这种结构。使用来自Small World of Words数据集的单词联想数据,我们比较了一对网络和一个包含N=3586个概念/节点的超图。在(1)只有心理语言学特征、(2)基于成对特征的特征汇总和(3)基于超图的汇总的可解释人工智能模型上进行了训练,结果显示高阶和特征丰富的超图模型包含比成对网络更丰富的信息,从而提高了单词具体性的预测能力。讨论了与以前关于联想知识和人类记忆中的概念聚类和分室化的研究的关系。