ISSN 0439-755X
CN 11-1911/B
主办:中国心理学会
   中国科学院心理研究所
出版:科学出版社

心理学报 ›› 2026, Vol. 58 ›› Issue (4): 590-602.doi: 10.3724/SP.J.1041.2026.0590 cstr: 32110.14.2026.0590

• 研究报告 • 上一篇    下一篇

单通道和多通道下的统计学习跨通道迁移

唐溢1, 赵亚军2, 曾清樟3, 张智君3, 吴圣楠1   

  1. 1重庆行政学院, 重庆 400041;
    2西南民族大学教育学与心理学院, 成都 610041;
    3浙江大学心理与行为科学系, 杭州 310022
  • 收稿日期:2025-03-17 发布日期:2026-01-16
  • 通讯作者: 唐溢, E-mail: tangy436@126.com
  • 基金资助:
    国家社会科学基金项目(21XDJ002)

Cross-modal transfer of statistical learning under unimodal and multimodal learning conditions

TANG Yi1, ZHAO Yajun2, ZENG Qingzhang3, ZHANG Zhijun3, WU Shengnan1   

  1. 1Chongqing Academy of Governance, Chongqing 400041, China;
    2College of Sociology and Psychology, Southwest University for Nationalities, Chengdu 610041, China;
    3Department of Psychology and Behavioral Sciences, Zhejiang University, Hangzhou 310022, China
  • Received:2025-03-17 Online:2026-01-16

摘要: 统计学习是人类从环境中提取规律信息的关键认知能力, 其跨通道迁移的特性与机制仍存理论争议。本研究通过4个实验系统考察统计学习在视听通道间的迁移机制。实验1采用经典统计学习范式, 验证了个体基于动物图片的视觉统计学习能力; 实验2在单一视觉通道学习后, 分别以动物图片和动物声音进行测试, 发现两种条件下统计学习效果无显著差异, 表明存在从视觉到听觉的跨通道迁移; 实验3通过在视听双通道同步呈现动物图片与无意义音节刺激流, 发现无论听觉刺激流是否具有统计规律, 视觉向听觉的统计学习迁移效应均稳定存在; 实验4进一步在视听双通道同步呈现动物声音与无意义图形刺激流, 结果显示从听觉向视觉的跨通道迁移亦显著。本研究综合表明, 统计学习具备稳定的跨通道双向迁移能力, 支持统计学习具有通道一般性的理论观点, 存在多层平行的统计结构表征系统可能是其认知基础。

关键词: 统计学习, 动物图片, 动物声音, 多通道, 跨通道迁移

Abstract: Statistical learning (SL), defined as an unconscious and automatic ability to extract regularities from the environment, has been shown to operate across multiple sensory modalities, including vision, audition, and touch. Although SL exhibits a certain degree of modality independence, these processes are not entirely isolated and may interact. From the perspective of object-feature processing, Frost et al. (2015) proposed the abstract rule representation hypothesis, suggesting that individuals may rely on four types of characteristics when learning inter-object regularities: modality specificity, stimulus specificity, modality generality, and SL jointly modulated by both modality and stimulus. However, theoretical disagreement remains regarding whether statistical regularities learned in one modality can be directly expressed in another—namely, whether cross-modal transfer occurs. Existing research on SL transfer has mainly focused on two areas: (a) transfer between low-level features within the same modality (e.g., from shape to color), which has not been extended to the cross-modal level, and (b) transfer between objects with semantic information, which has also not addressed cross-modal processing mechanisms. Against this backdrop, the present study used animal pictures and animal sounds as materials to examine the cross-modal transfer of SL between the visual and auditory modalities.
This study included four experiments that integrated the cross-modal transfer and multimodal SL paradigms to investigate cross-modal transfer of SL in realistic object contexts systematically. Experiment 1 constructed a visual stimulus stream using animal pictures to verify visual SL. Experiment 2 employed a cross-modal transfer paradigm in which participants were visually familiarized only with animal pictures and then tested with either animal pictures or animal sounds. By comparing performance between visual-visual and visual-auditory conditions, the experiment evaluated whether visual SL transfers across different modalities. Experiment 3 used a multimodal learning approach to separate modality-specific learning from cross-modal transfer. It aimed to (a) examine whether SL in the visual modality is independent of that in the auditory modality, and (b) investigate the relationship between visual-to-auditory transfer and auditory SL under multimodal learning conditions. Experiment 4 assessed the transfer of SL from audition to vision and, together with Experiment 3, examined the bidirectionality of cross-modal transfer.
Results showed that Experiment 1 successfully validated visual SL with animal pictures, confirming previous findings (e.g., Otsuka et al., 2013). In Experiment 2, learned statistical regularities through visual unimodal exposure persisted within the visual modality and also transferred to the auditory modality, indicating comparable learning effects across both senses. Experiment 3 revealed that multimodal input did not significantly interfere with unimodal visual or auditory SL, aligning with studies by Li et al. (2018) and Mitchel and Weiss (2011), and supporting the idea that SL operates relatively independently across sensory modalities. Furthermore, regardless of whether the auditory stream contained statistical regularities, the visual-to-auditory transfer effect remained robust, suggesting that cross-modal transfer can occur alongside unimodal SL. Experiment 4 confirmed that statistical regularities learned through audition could transfer to vision. Together, these experiments offer converging evidence for bidirectional cross-modal transfer of SL, indicating that it is not modality-specific but instead reflects a general cognitive mechanism.
In summary, the study presents three main conclusions: (1) SL of real animal objects shows bidirectional cross-modal transfer; (2) SL in visual and auditory modalities is fairly independent; and (3) unimodal SL and cross-modal transfer can occur independently in parallel and simultaneously, supporting the idea of a multilevel statistical representation system.

Key words: statistical learning, animal picture, animal sound, multimodal, cross-modal transfer