ISSN 0439-755X
CN 11-1911/B

Acta Psychologica Sinica ›› 2025, Vol. 57 ›› Issue (6): 929-946.doi: 10.3724/SP.J.1041.2025.0929

• Academic Papers of the 27 th Annual Meeting of the China Association for Science and Techn •     Next Articles

When AI “possesses” personality: Roles of good and evil personalities influence moral judgment in large language models

JIAO Liying1(), LI Chang-Jin2, CHEN Zhen2, XU Hengbin2, XU Yan2()   

  1. 1Department of Psychology, School of Humanities and Social Sciences, Beijing Forestry University, Beijing 100083, China
    2Beijing Key Laboratory of Applied Experimental Psychology, National Demonstration Center for Experimental Psychology Education (Beijing Normal University), Faculty of Psychology, Beijing Normal University, Beijing 100875, China
  • Published:2025-06-25 Online:2025-04-15
  • Contact: JIAO Liying,XU Yan E-mail:jiaoliying316@163.com;xuyan@bnu.edu.cn

Abstract:

At the intersection of technology and morality, a critical question arises: Can large language models (LLMs) simulate good and evil personalities, and does this capacity influence their performance in moral judgment tasks? This study investigated the moral judgment characteristics of LLMs when simulating different good and evil personalities, as well as the similarities and differences between these patterns and those of humans. Across two studies, we analyzed moral judgment data generated by two LLMs—ERNIE 4.0 and GPT-4 (N = 4,832)—alongside responses from human participants (N = 370). The results revealed that: (1) LLMs are capable of successfully simulating varying levels of good and evil personalities; (2) the personality configuration significantly affects the moral judgments made by LLMs; and (3) a personality hierarchy emerges in the alignment between human and LLMs’ responses: good personality plays a more critical role than evil personality (inter-personality hierarchy), and within the good personality, conscientiousness and integrity dimension exerts the strongest influence (intra-personality hierarchy). This research constructed a theoretical model of good and evil personalities in LLMs under moral judgment tasks, contributing to a deeper understanding of how simulated personalities function in AI moral reasoning. The findings provided a theoretical foundation for promoting moral alignment in artificial intelligence systems.

Key words: Large Language Models, good and evil personalities, moral judgment, human-AI consistency, personality hierarchy