ISSN 1671-3710
CN 11-4766/R
主办:中国科学院心理研究所
出版:科学出版社

心理科学进展 ›› 2025, Vol. 33 ›› Issue (5): 863-886.doi: 10.3724/SP.J.1042.2025.0863

• 研究前沿 • 上一篇    

字母位置编码的模型对比及其效应解释

李璜夏1, 陈新炜2, 药盼盼1   

  1. 1北京语言大学心理学院, 北京100083;
    2北京外国语大学中国语言文学学院, 北京100089
  • 收稿日期:2024-12-17 出版日期:2025-05-15 发布日期:2025-03-20

Comparison of models for letter position encoding and their explanations of experimental effects

LI Huangxia1, CHEN Xinwei2, YAO Panpan1   

  1. 1School of Psychology, Beijing Language and Culture University, Beijing 100083, China;
    2School of Chinese Language and Literature, Beijing Foreign Studies University, Beijing 100089, China
  • Received:2024-12-17 Online:2025-05-15 Published:2025-03-20

摘要: 在视觉词汇识别过程中, 字母的位置信息发挥了重要的作用。过去几十年间, 关于字母位置编码的研究极大地推动了各种理论框架的发展, 这些理论旨在解释不同的实验效应及其背后的认知加工机制。文章系统介绍了关于字母位置编码的6个理论模型, 包括重叠模型(the Overlap Model)、开放双字母组模型(the Open-Bigram Model)、序列编码模型(the SERIOL Model)、空间编码模型(the Spatial Coding Model)、贝叶斯读者模型(the Bayesian Reader)以及N-字母组位置编码模型(PONG: the Positional Ordering of N-Grams)。这些模型涵盖了从重叠编码到序列和空间编码等不同的认知加工机制, 代表了字母位置编码领域中的重要理论框架。文章从模型结构、理论基础、词汇识别逻辑、跨语言适应性解释以及常见效应解释等方面进行对比分析, 并且对模型尚未能解释的效应进行了总结。基于对这些模型的分析总结, 未来模型建构可以整合更多实证研究结果以及不同类型的实验数据, 以增强模型解释力度。此外, 考虑到跨语言因素以及第二语言的研究成果, 探究字母位置加工及相关模型的跨语言一致性将是一个有价值的研究方向。

关键词: 字母位置编码模型, 词汇识别, 转置效应, 模型对比

Abstract: This article compares and analyzes six classic letter position encoding models, exploring the role of letter position in visual word recognition and the underlying cognitive mechanisms. The models discussed include the Overlap Model, the Open Bigrams Model, the Sequential Encoding Regulated by Inputs to Oscillations within Letter Units (the SERIOL Model), the Spatial-coding Model, the Bayesian Reader, and the Positional Ordering of N-Grams (the PONG). Each model is grounded on a distinct cognitive theory, offering unique perspectives on how letter positions are encoded and simulating various effects observed in word recognition.
A common feature of these models is the claim that letter position plays a crucial role in word recognition. All models effectively account for several experimental effects, such as letter transposition effects and the effects of letter repetition, insertion, and deletion. Nonetheless, significant differences emerge in these models’ theoretical frameworks and conceptualizations of letter position encoding. The Overlap Model and the Bayesian Reader suggest that letter position encoding is highly susceptible to noise. The Overlap Model tolerates position shifts through an overlapping encoding mechanism, while the Bayesian Reader interprets position adjustment through a probabilistic framework, dynamically adjusting based on context to fit various cognitive scenarios. In contrast, the SERIOL Model encodes letter position via temporal activation patterns and multi-level structures (e.g., retinal layer, feature layer, letter layer), emphasizing the sequential and temporal characteristics of letter positions. The Spatial Encoding Model employs spatial phase coding and a two-dimensional coordinate system to process letter positions, enhancing encoding precision through external letter banks and word length modules. The Open Bigrams Model proposes a flexible encoding mechanism, focusing on the relative positions of adjacent letter pairs rather than strict letter order, thus offering a moderately flexible approach. The PONG introduces a theory of N-grams position encoding, positing that the position of N-grams is more critical than individual letter positions.
Consequently, these models differ in their explanations of certain experimental effects of letter transposition. The Open Bigrams Model suggests that transposed nonwords overlap more with base words in terms of letter pairs, thereby triggering a stronger transposition effect. The Overlap Model posits that letter representations are distributed across sequential positions. Thus transposed nonwords, having more overlap with base words’ position distribution, are recognized more quickly than substituted nonwords. The Bayesian Reader, calculates the probability of each letter’s position and views transposed letters as having a smaller edit distance from the target word, thus enhancing the match score. The SERIOL Model incorporates the Open Bigrams framework and argues that the transposition conditions activate more matching letter pairs than the substitute conditions, further strengthening the transposition effect. The Spatial Encoding Model emphasizes phase encoding of letter positions, suggesting that transposed letters preserve more stable relative phase relationships, thus facilitating better word matching. Meanwhile, the PONG activates matching N-grams and contends that transposed nonwords engage more appropriate N-grams with similar hemispheric lateralization to the target word, resulting in better recognition compared to substituted nonwords.
Additionally, each model offers unique insights into specific phenomena. The PONG Model, for example, explains the flanker effect using N-grams, a phenomenon has not been addressed by other models. Both the SERIOL and PONG Models account for optimal fixation locations through hemispheric lateralization. The Open Bigrams Model explains the influence of reading direction by positing that letter detectors are based on the relative position of eye fixations along the horizontal line.
Based on the analysis and discussions of the above models, there are several key directions for the further development of letter position encoding models. First, there is a need to expand the scope of models to account for factors like prediction, word frequency, and transposition across word boundaries, which remain insufficiently explored. Second, integrating word recognition models with eye movement control models could improve ecological validity in understanding letter position encoding in reading. Third, research into Chinese character position processing should be further explored. Fourth, more attention should be paid to cross-language differences and the differential experimental effects observed across languages. Finally, with advances in EEG and neuroimaging technologies, integrating diverse sources of data would optimize existing models and address theoretical gaps.

Key words: computational models of letter position encoding, word recognition, transposition effect, model comparison

中图分类号: