ISSN 0439-755X
CN 11-1911/B

Acta Psychologica Sinica ›› 2026, Vol. 58 ›› Issue (3): 399-415.doi: 10.3724/SP.J.1041.2026.0399

• Academic Papers of the 27th Annual Meeting of the China Association for Science and Technology • Previous Articles     Next Articles

LLMs Amplify Gendered Empathy Stereotypes and Influence Major and Career Recommendations*

DAI Yiqing1, MA Xinming2, WU Zhen1,3()   

  1. 1Department of Psychological and Cognitive Sciences, Tsinghua University, Beijing 100084, China
    2Faculty of Education, Beijing Normal University, Beijing 100875, China
    3Lab for Lifelong Learning, Tsinghua University, Beijing 100084, China
  • Published:2026-03-25 Online:2025-12-26
  • Contact: WU Zhen E-mail:zhen-wu@mail.tsinghua.edu.cn
  • Supported by:
    National Natural Science Foundation of China(32271110);National Natural Science Foundation of China(62441614);Tsinghua University Initiative Scientific Research Program(20235080047)

Abstract:

Large language models (LLMs) are increasingly deployed in highly sensitive domains such as education and career guidance, raising concerns about their potential to reproduce and amplify social biases. The present research examined whether LLMs exhibit gendered empathy stereotypes—specifically, the belief that “women are more empathetic than men”—and whether such stereotypes influence downstream recommendations. Three studies were conducted. Study 1 compared LLMs with human participants and found that across six leading LLMs, gendered empathy stereotypes were significantly stronger than those observed in humans across three facets of empathy: emotional empathy, attention to others’ feelings, and behavioral empathy. Study 2 manipulated input language (Chinese vs. English) and gender-identity priming (male vs. female), demonstrating that English prompts and female priming elicited stronger gendered empathy stereotypes. Study 3 focused on major and career recommendation tasks and revealed that LLMs systematically recommended high-empathy majors and professions to women, while directing men toward low-empathy fields. Together, these findings indicate that LLMs exhibit pronounced gendered empathy stereotypes, that these biases vary across input context, and that they can transfer into real-world recommendation scenarios. This research offers theoretical insights into bias formation in LLMs and provides practical implications for improving fairness in AI systems used in educational and career guidance.

Key words: large language models, gender stereotypes, empathy, AI recommendations, human-computer interaction