ISSN 0439-755X
CN 11-1911/B
主办:中国心理学会
   中国科学院心理研究所
出版:科学出版社

心理学报 ›› 2026, Vol. 58 ›› Issue (2): 308-322.doi: 10.3724/SP.J.1041.2026.0308 cstr: 32110.14.2026.0308

• 研究报告 • 上一篇    下一篇

“零样本语言学习”:大语言模型能“像人一样”习得语境中的情感吗?

吴诗玉, 王亦赟   

  1. 上海交通大学外国语学院, 上海 200240
  • 收稿日期:2025-01-27 发布日期:2025-12-03 出版日期:2026-02-25
  • 通讯作者: 吴诗玉, E-mail: shiyuw@sjtu.edu.cn

Zero-shot language learning: Can large language models (LLMs) acquire contextual emotion in a human-like manner?

WU Shiyu, WANG Yiyun   

  1. School of Foreign Languages, Shanghai Jiao Tong University, Shanghai 200240, China
  • Received:2025-01-27 Online:2025-12-03 Published:2026-02-25

摘要: 本研究旨在检验大语言模型(LLMs)能否在“零样本”条件下通过阅读附带习得单词所出现的语境情感, 并评估情感效价与语境变异性对词汇学习的影响。研究采用跨模型-人类对比, 4种LLMs与3组学习者在统一材料中学习嵌入不同情感(积极/中性/消极)与重复/变化语境的目标词, 并以多项测试衡量情感迁移及词形、词义习得效果。结果显示, LLMs与人类模式一致, 能将语境情感迁移至目标词, 并在语言生成中保持情感一致; 而且也呈现“积极情感优势”“语境变异优势”, 且在定义生成中出现语境情感与语境变异的交互效应。文章提出“双重机制框架”, 认为LLMs在功能层面具备类人的情感语义学习能力, 但其机制基于统计共现与向量优化, 异于人类的具身与社会加工。本研究为情感计算、人机交互伦理与词汇教学提供启示。

关键词: 大语言模型, 零样本学习, 情感学习

Abstract: Emotion is a structural resource in human cognition that guides attention, memory, and social coordination. During incidental vocabulary acquisition (IVA), readers often internalize the affective tone of surrounding discourse and transfer it to novel words (“contextual emotion transfer/semantic prosody”). Recent LLMs appear to display analogous behavior despite lacking embodiment, raising the question of whether they can acquire contextual emotion in a human-like manner and whether the same contextual factors shape both human and model learning. Building on usage-based and distributional accounts, we expected two robust regularities to hold across agents: a positivity advantage (higher contextual valence predicts better learning) and a variability advantage (varied contexts outperform repeated ones). We further hypothesized that, in more demanding recall (definition generation), contextual valence would interact with variability, such that positive emotion would amplify the benefits of varied contexts.
We conducted zero-shot, parallel evaluations with four representative LLMs (Ernie Bot 3.5, ChatGPT/GPT-4, Gemini 1.5 Pro, LLaMA 3.1-8B) and three human cohorts matched to prior IVA paradigms (English L1, Chinese L1, English L2; 306 participants). Each agent learned nine pseudowords embedded in 45 two-sentence texts spanning positive, neutral, and negative contexts; context variability was manipulated between repeated versus varied exposures. After reading, LLMs completed (a) valence rating and sentence production (emotion transfer) and (b) orthographic choice, definition matching, and definition generation (form/meaning). LLMs were evaluated in strictly isolated zero-shot sessions with no task-specific supervision or fine-tuning. Ordinal mixed-effects models (CLMM) analyzed ratings; linear/logistic mixed-effects models analyzed production and accuracy, with random effects for participant/LLM session, item, and denotation class.
Contextual emotion transferred reliably to targets: across humans and LLMs, ratings followed positive > neutral > negative, and generated sentences aligned in polarity with the learning context. For vocabulary learning, both groups exhibited a positivity advantage—higher contextual valence significantly predicted better meaning performance—and a variability advantage—varied contexts significantly outperformed repeated contexts in definition matching and definition generation. In recall, valence interacted with variability: positive emotion amplified gains under varied exposure for both humans and LLMs, yielding the largest improvements in definition generation. LLMs frequently matched or exceeded human accuracy in form recognition and often reached higher overall accuracy on meaning tasks while preserving the same qualitative patterns. These effects held in mixed-effects analyses controlling for participant/session, item, and denotation, and were observed without providing LLMs with examples, feedback, or fine-tuning.
The study showed that LLMs did acquire contextual emotion and reproduced core human regularities (positivity and variability advantages; valence-by-variability interaction in recall). We interpret the convergence via a Dual-Mechanism perspective: human emotion learning is embodied and socially situated, whereas LLM “emotion” arises from distributional co-occurrence and vector-space optimization; distinct mechanisms can yield functionally similar behavior. The findings advance theories of emotion-language interaction and support context variability as a general driver of vocabulary learning. Practically, emotion-sensitive LLM behavior can enhance educational and communicative applications, while necessitating safeguards against unintended amplification of corpus-borne affective biases.

Key words: large language models, zero-shot learning, emotion learning

中图分类号: