School of Psychological Sciences, The University of Melbourne.
Department of Cognitive Sciences, Hanyang University.
Cogn Sci. 2024 Sep;48(9):e13494. doi: 10.1111/cogs.13494.
Models of word meaning that exploit patterns of word usage across large text corpora to capture semantic relations, like the topic model and word2vec, condense word-by-context co-occurrence statistics to induce representations that organize words along semantically relevant dimensions (e.g., synonymy, antonymy, hyponymy, etc.). However, their reliance on latent representations leaves them vulnerable to interference, makes them slow learners, and commits to a dual-systems account of episodic and semantic memory. We show how it is possible to construct the meaning of words online during retrieval to avoid these limitations. We implement a spreading activation account of word meaning in an associative net, a one-layer highly recurrent network of associations, called a Dynamic-Eigen-Net, that we developed to address the limitations of earlier variants of associative nets when scaling up to deal with unstructured input domains like natural language text. We show that spreading activation using a one-hot coded Dynamic-Eigen-Net outperforms the topic model and reaches similar levels of performance as word2vec when predicting human free associations and word similarity ratings. Latent Semantic Analysis vectors reached similar levels of performance when constructed by applying dimensionality reduction to the Shifted Positive Pointwise Mutual Information but showed poorer predictability for free associations when using an entropy-based normalization. An analysis of the rate at which the Dynamic-Eigen-Net reaches asymptotic performance shows that it learns faster than word2vec. We argue in favor of the Dynamic-Eigen-Net as a fast learner, with a single-store, that is not subject to catastrophic interference. We present it as an alternative to instance models when delegating the induction of latent relationships to process assumptions instead of assumptions about representation.
利用大型文本语料库中的词汇使用模式来捕捉语义关系的词汇意义模型,如主题模型和 word2vec,将词与上下文共现的统计数据压缩为表示形式,从而沿着语义相关维度组织词汇(例如同义词、反义词、上下位词等)。然而,它们对潜在表示的依赖使其容易受到干扰,使它们成为缓慢的学习者,并承诺对情节和语义记忆进行双重系统解释。我们展示了如何在检索过程中在线构建单词的含义,以避免这些限制。我们在关联网络中实现了词汇意义的扩展激活,关联网络是一种单层的高度递归关联网络,称为动态特征网络,我们开发了它来解决在扩展到处理非结构化输入领域(如自然语言文本)时,关联网络的早期变体所面临的限制。我们表明,使用独热编码的动态特征网络进行扩展激活在预测人类自由联想和单词相似性评分方面优于主题模型,并达到与 word2vec 相似的性能水平。通过将降维应用于移位正点互信息来构建的潜在语义分析向量在构造相似水平的性能时,当使用基于熵的归一化时,其对自由联想的预测能力较差。对动态特征网络达到渐近性能的速度的分析表明,它比 word2vec 学习更快。我们赞成动态特征网络作为一种快速学习者,具有单一存储,不会受到灾难性干扰。当将潜在关系的归纳委托给过程假设而不是表示假设时,我们将其作为实例模型的替代方案。