Loading...
ISSN 0439-755X
CN 11-1911/B
主办:中国心理学会
   中国科学院心理研究所
出版:科学出版社

Archive

    For Selected: Toggle Thumbnails
    The special issue on “Artificial Intelligence Psychology and Governance”
    Human advantages and psychological transformations in the era of artificial intelligence
    WU Michael Shengtao, PENG Kaiping
    2025, 57 (11):  1879-1884.  doi: 10.3724/SP.J.1041.2025.1879
    Abstract ( 2960 )   HTML ( 198 )  
    PDF (413KB) ( 4386 )  

    In the era of artificial intelligence (AI), the boundaries between humans and machines have become blurred, and re-understanding and developing humanity's unique advantages are increasingly prominent and urgent. Meanwhile, with the rapid development of technology and scientific paradigm, a broad psychology encompassing the minds and behaviors of humans, animals, and machines is emerging. Recent researchers have conducted a series of studies on the psychology and governance of AI, from the perspectives of impacts of AI, new human-machine relationships, AI methods, and interdisciplinary empowerment. Future psychology researchers should focus on human society and future development, and reflect on the status of humanity and human dignity under the impact of AI, especially the unique advantages derived from human evolution as well as the expansions of human nature and identity; truly master and utilize AI technologies to empower the development of psychology, making mind research on the black box of human consciousness and complex social behavior more precise and efficient, and promoting AI-based mind computation and intervention across time and space scales and personalized interventions. More important, they must consider how psychology (with strengths in studying human nature, social relations, and ethical values) could empower the development of AI, by exploring AI cognition and its comparison with humans and animals, which is critical for promoting the AI application and governance in a human-machine symbiotic society.

    References | Related Articles | Metrics
    Human-AI cooperation makes individuals more risk seeking: The mediating role of perceived agentic responsibility
    GENG Xiaowei, LIU Chao, SU Li, HAN Bingxue, ZHANG Qiaoming, WU Mingzheng
    2025, 57 (11):  1885-1900.  doi: 10.3724/SP.J.1041.2025.1885
    Abstract ( 3240 )   HTML ( 266 )  
    PDF (766KB) ( 4645 )   Peer Review Comments

    Risk decision-making involves choices made by individuals when they are uncertain about future outcomes. With advancements in artificial intelligence (AI), AI can now assist humans in making decisions. For instance, human drivers and AI drivers cooperate to carry out driving tasks, human doctors and AI doctors can collaborate on medical decisions. Currently, it is unclear how AI affects individuals’ risk decision-making during such collaborations, which is crucial for enhancing the quality of human-AI decision-making. Therefore, studying the impact of human-AI cooperation on individuals’ risk decision-making is essential.

    In Experiment 1a, a total of 100 participants were recruited from one university. Employing a within-subject design, the independent variable was the partner type (i.e., human-human cooperation, human-AI cooperation, or no partner), while the dependent variable measured individuals’ risk decision-making using the Balloon Analogue Risk Task (BART). In Experiment 1b, a total of 151 participants were recruited from another university and randomly assigned to two conditions: human-human cooperation and human-AI cooperation. As in Experiment 1a, the dependent variable remained the same. To investigate the mediating role of individual agentic responsibility, Experiment 2 recruited 199 participants from a university. This experiment utilized a between-subjects design, with the independent variable being the partner type (i.e., human-human cooperation or human-robot cooperation). Individual agentic responsibility was assessed by measuring the extent to which participants assumed responsibility for their tasks, and the dependent variable was individuals’ risk levels as measured by the BART. Experiment 3 further explored the moderating effect of outcome feedback. Participants received feedback based on their BART performance in Experiment 2, categorized as success or failure, and then assessed their perceived agentic responsibility before completing the BART again.

    The results of Experiment 1a and 1b showed that participants in the control group (i.e., without cooperation) exhibited the highest risk-taking behavior, while those engaged in human-AI cooperation took greater risks than those in human-human cooperation. Results from Experiment 2 demonstrated that individual agentic responsibility partially mediated the effect of human-AI cooperation on individuals’ risk decision-making. Specifically, participants reported a higher sense of agentic responsibility in human-AI cooperation compared to human-human cooperation, which contributed to increased risk-taking. Experiment 3 revealed that outcome feedback significantly moderates the mediating role of individual agentic responsibility regarding the influence of human-AI cooperation (versus human-human cooperation) on individuals’ risk decision-making. Notably, under success conditions, participants attributed greater responsibility to themselves in human-AI collaboration compared to human-human collaboration. Conversely, under failure conditions, there was no significant difference in responsibility attribution between the two types of collaboration.

    This research demonstrates that collaboration with AI can enhance an individual's propensity for risk-taking. Moreover, the influence of human-AI cooperation, compared to human-human cooperation, on individuals’ risk decision-making is mediated by a sense of individual agentic responsibility and moderated by outcome feedback. These findings offer significant theoretical insights. Furthermore, this study holds substantial practical implications by aiding individuals in understanding how collaboration with AI impacts their risk-taking behaviors.

    Figures and Tables | References | Related Articles | Metrics
    Unity without uniformity: Humans’ social creativity strategy under generative artificial intelligence salience
    ZHOU Xiang, BAI Boren, ZHANG Jingjing, LIU Shanrou
    2025, 57 (11):  1901-1913.  doi: 10.3724/SP.J.1041.2025.1901
    Abstract ( 1818 )   HTML ( 114 )  
    PDF (1048KB) ( 2306 )   Peer Review Comments

    The rapidly advancing artificial intelligence (AI) technology has sparked concerns akin to opening a Pandora’s box. In recent years, the remarkable progress of generative AI has further amplified this uncertainty, bringing forth numerous challenges to human society. One such challenge is the symbolic threat to human distinctiveness posed by generative AI. This threat is difficult to avoid and may generate widespread and intricate negative impacts. Therefore, grounded in social identity theory, this study examines individuals’ use of social creativity strategies in response to generative AI salience, while also investigating the moderating role of anthropocentrism in this process. To investigate these effects, this research conducted four progressive studies. Study 1 used two survey rounds to identify mind dimensions perceived as overlapping or distinct between humans and generative AI, revealing the specific forms of social creativity strategies in response to generative AI salience. Study 2 employed a single-factor, three-level between-subjects design to demonstrate that generative AI salience leads individuals to adopt social creativity strategies, particularly by emphasizing the importance of distinguishing dimensions. Study 3 introduced anthropocentrism as a moderating variable using a 2×2 between-subjects design and found that anthropocentrism strengthens the use of social creativity strategies when generative AI is salient. Finally, Study 4 controlled for the effects of social competition strategies by presenting participants with both social creativity and social competition scenarios, further demonstrating that generative AI salience leads individuals to adopt social creativity strategies, and anthropocentrism enhances social creativity strategy use in generative AI salient contexts. The findings reveal several key insights across four studies. Study 1 identified the specific mind dimensions that individuals perceive as overlapping or distinct between human and generative AI, demonstrating that people actively redefine these distinctions when faced with generative AI salience. Study 2 showed that individuals respond to generative AI salience by employing social creativity strategies, specifically by emphasizing the importance of distinguishing dimensions of the human mind. Study 3 introduced anthropocentrism as a moderating factor and demonstrated that individuals with higher anthropocentric beliefs are more likely to adopt social creativity strategies when generative AI is salient. Finally, Study 4 controlled for the influence of social competition strategies and provided additional evidence that generative AI salience leads individuals to adopt social creativity strategies, and anthropocentrism enhances the use of social creativity strategies. In summary, this study demonstrates that generative AI salience prompts the application of social creativity strategy, with a moderating role played by individuals’ anthropocentrism. The findings reveal the specific manifestations and positive implications of individuals’ use of social creativity strategies in contexts where generative AI is made salient. They also elucidate the functional role of anthropocentrism and offer new insights into how social creativity strategies can be leveraged to promote harmonious human−AI coexistence in the era of intelligence.

    Figures and Tables | References | Related Articles | Metrics
    When design meets AI: The impact of AI design products on consumers’ response patterns
    LI Bin, RUI Jianxi, YU Weinan, LI Aimei, YE Maolin
    2025, 57 (11):  1914-1932.  doi: 10.3724/SP.J.1041.2025.1914
    Abstract ( 2079 )   HTML ( 159 )  
    PDF (6604KB) ( 4656 )   Peer Review Comments

    In recent years, the prevalence and popularity of artificial intelligence (AI) products and services have increased significantly. While the advantages of utilizing AI applications, including cost-effectiveness, standardization, and efficiency, are widely recognized, customer attitudes towards AI involvement are polarized. Prior research has suggested that consumer resistance or acceptance of AI can be influenced by various factors, such as medical services, product recommendations, computer services, and more. However, limited studies have examined AI's impact on creative tasks, such as product design. Therefore, the objective of this study is to investigate consumer preferences for the designer role in different product categories and to identify potential mechanisms and boundary conditions that may vary.

    To attain the objectives of this research, we conducted one pre-study and six studies that focused on commonplace products in daily life. Study 1A (N = 181) adopted a 2 (product type: nostalgia vs. innovation) × 2 (designer role: human design vs. AI design) design and measured participants' attitudes towards the products. Study 1B (N = 258) utilized a between-subject design of 2 (product type: nostalgia vs. innovation) × 3 (designer role: human vs. AI vs. human-AI collaborative design). In Study 2A, we considered the potential for human-machine collaboration, which expands the design resources and examined the mediating impact of processing fluency. Specifically, Study 2A (N = 304) adopted a between-subject design of 2 (product type: nostalgia vs. innovation) × 4 (designer role: human design vs. human-led/AI-assisted design vs. AI-led/human-assisted design vs. AI design). Study 2B (N = 167) conducted a mixed design of 2 (product type: nostalgia vs. innovation) × 3 (designer role: human vs. AI design). Studies 3 (N = 218) and 4 (N = 290) investigated the moderating effects of AI characteristics (AI anthropomorphism) and consumer characteristics (self-constructed types), respectively. Specifically, Study 3 manipulated the degree of AI anthropomorphism of the designer, while Study 4 manipulated the type of self-construction of the participants.

    Our hypotheses were confirmed, as this study reveals a significant interaction between product types and designer roles. Specifically, we found a congruence effect between “human design-nostalgic products” and “AI design-innovative products”. The robustness of our findings is demonstrated by the diversity of stimuli and samples employed in this research. Additionally, we discovered that processing fluency acts as a mediating mechanism that explains how the interaction between product type and designer role influences consumer responses. In other words, consumers express more negative reactions towards nostalgic products designed by AI due to their perception of lower processing fluency. Finally, our results emphasize the moderating effects of AI-human collaborative design scenarios, AI anthropomorphic characteristics, and consumer self-construction types.

    From a social psychological perspective, this study contributes to the literature on AI aversion and product design by examining people's behavioral tendencies towards AI design. Specifically, we focused on consumer attitudes towards the designer role behind different product types. This investigation helps to enhance our understanding of how people's aversion to algorithms applies to the design field. Our findings align with prior research on the human psychological perception of artificial intelligence, revealing the mechanism behind the matching effect of processing fluency. Furthermore, our study offers additional evidence that AI personification and the type of consumer self-construction significantly impact consumer responses.

    Figures and Tables | References | Related Articles | Metrics
    Impact of trusting humanoid intelligent robots on employees’ job dedication intentions: An investigation based on the classification of human−robot trust
    TANG Xiaofei, WANG Changmei, SUN Xiaodong, CHANG En-Chung
    2025, 57 (11):  1933-1950.  doi: 10.3724/SP.J.1041.2025.1933
    Abstract ( 1156 )   HTML ( 69 )  
    PDF (1936KB) ( 1636 )   Peer Review Comments

    As humanoid intelligent robots (HIRs) become increasingly integrated into organizations to provide emotional and functional support, understanding and fostering human−robot trust has become a critical area of focus. This study explores the formation and impact of human−robot trust from the perspective of the Unique Agent Hypothesis. We propose that human−robot trust comprises two distinct dimensions: emotional repair trust and functional aiding trust. Among various forms of human−robot collaboration, self-repair and friendship-repair forms primarily trigger emotional repair trust, while intelligence-aiding and physical-aiding forms more effectively enhance functional aiding trust. Furthermore, employees who develop emotional repair trust (vs. functional aiding trust) in HIRs perceive there to be greater organizational warmth, which in turn strengthens their job dedication intentions. Conversely, employees who develop functional aiding trust (vs. emotional repair trust) in HIRs perceive there to be higher organizational competence, thereby enhancing their job dedication intentions. In addition, interaction orientation and task orientation are introduced as crucial situational moderators. This research included one qualitative study and three quantitative studies. Study 1 (qualitative and experimental study) identified the two dimensions of human−robot trust and investigated the relationship between four forms of human−robot collaboration and trust. Study 2 (experimental study) verified how human−robot trust affects employees’ job dedication intentions through perceived organizational competence and warmth. Study 3 (experimental study) tested how interaction and task orientations moderate these effects. The main findings of this study are as follows. Study 1 provided evidence that emotional repair trust and functional aiding trust are two distinct dimensions of human−robot trust. Results showed that both self-repair and friendship-repair forms of human−robot collaboration elicited higher emotional repair trust than functional aiding trust, whereas both intelligence-aiding and physical-aiding forms led to higher functional aiding trust than emotional repair trust. Study 2 further demonstrated that functional aiding trust led to higher perceived organizational competence but lower perceived organizational warmth than did emotional repair trust. Further analysis (coded as 0 = emotional repair trust and 1 = functional aiding trust) in Study 2 revealed that the indirect effect of perceived organizational warmth was significantly negative, while the indirect effect of perceived organizational competence was significantly positive. Study 3 showed that interaction orientation (task orientation) positively (negatively) moderated the relationship between human−robot trust (coded as 0 = functional aiding trust and 1 = emotional repair trust) and perceived organizational warmth while negatively (positively) moderating the relationship between human−robot trust and perceived organizational competence. This study contributes to the emerging field of machine behavior by embedding HIR into the network of internal organizational dynamics and examining the impact of human−robot trust on employees’ job dedication intentions. The findings of this study suggest that employees’ trust in HIR agents’ emotional and functional support enhances perceived organizational warmth and competence, promoting their job dedication intentions. This demonstrates that organizational support provided by HIR agents can generate positive effects, thereby enriching and expanding the existing framework of organizational support theory.

    Figures and Tables | References | Related Articles | Metrics
    Safety trust in intelligent domestic robots: Human and AI perspectives on trust and relevant influencing factors
    YOU Shanshan, QI Yue, CHEN JunTing, LUO Lei, ZHANG Kan
    2025, 57 (11):  1951-1972.  doi: 10.3724/SP.J.1041.2025.1951
    Abstract ( 1295 )   HTML ( 80 )  
    PDF (3001KB) ( 1692 )   Peer Review Comments

    As a result of the rapid development of intelligent domestic robot technology, safety concerns have emerged as a new challenge in human‒robot trust dynamics. This study explores and validates novel critical dimensions of trust that influence human and AI users’ perceptions of intelligent domestic robots, with a particular focus on safety trust. The research involves three comprehensive studies, each of which addresses different aspects of these dimensions.

    In Study 1, we developed a safety trust scale pertaining specifically to intelligent domestic robots. This scale was rigorously tested to confirm the stability and validity of its three-dimensional structure, which included performance, relational, and safety trust. The scale’s psychometric properties were evaluated on the basis of factor analysis and reliability testing, thereby ensuring that it could accurately measure trust across different contexts and populations.

    Study 2 explored the static characteristics of robots, such as their anthropomorphism, their height, and the visibility of their embedded cameras. We revealed that human participants exhibited higher levels of safety trust toward robots that were shorter in height and had fewer conspicuous cameras. Interestingly, the degree of anthropomorphism was determined to play a significant role in determining participants’ sensitivity to these static features.

    Study 3 expanded the investigation to encompass the dynamic characteristics of robots, such as movement speed, interaction scenario and camera operation (i.e., turning the camera off). The results indicated that slower-moving robots were generally perceived as safer, and higher levels of safety trust were attributed to them. Moreover, the action of turning off a robot’s camera during interactions was observed to significantly enhance safety trust among human users. The study also highlighted the fact that the influence of these dynamic features varied across different interaction scenarios, thus suggesting that situational factors play crucial roles in shaping trust perceptions.

    Furthermore, a comparative analysis between human and AI users revealed a certain degree of consistency in safety trust judgments. Both human and AI users were generally aligned in terms of their trust assessments on the basis of both static and dynamic robot features. However, the AI’s sensitivity to the visibility of robot cameras was notably lower than that of humans, thus suggesting that AI may prioritize different factors in the context of assessing safety trust.

    Overall, the findings of this research provide valuable insights into the design and manufacturing of intelligent domestic robots, including by emphasizing the importance of considering both static and dynamic features in the process of enhancing safety trust. The results also offer theoretical and practical guidance for the development of trust models that can be applied in various intelligent home environments, thereby ultimately contributing to the advancement of human‒robot interactions.

    Figures and Tables | References | Related Articles | Metrics
    Perceived unsustainability decreases acceptance of artificial intelligence
    WEI Xinni, YU Feng, PENG Kaiping
    2025, 57 (11):  1973-1987.  doi: 10.3724/SP.J.1041.2025.1973
    Abstract ( 1089 )   HTML ( 68 )  
    PDF (534KB) ( 1281 )   Peer Review Comments

    Climate change and environmental issues significantly impact human health and well-being, posing substantial challenges to global sustainable development. Addressing these challenges necessitates both immediate and long-term solutions. While climate change itself does not inherently elicit a moral response, potentially hindering public engagement in climate action, artificial intelligence (AI)—encompassing robots, algorithms, and models—emerges as a promising ally. AI’s ability to learn from experience, adapt to new inputs, and perform human-like tasks positions it as a critical tool in addressing environmental challenges.

    However, AI presents a dual-edged impact on the environment. While it can support ecological governance and promote sustainability, its energy-intensive nature and associated carbon emissions may undermine the environment. This study investigates the environmental implications of AI, focusing on how perceptions of AI’s sustainability influence public acceptance and the underlying psychological mechanisms.

    Drawing on prior research, we hypothesized that individuals tend to avoid using unsustainable AI due to perceptions of diminished morality and heightened dependency (i.e., lower perceived agency). To test this, we conducted five studies employing diverse methodologies and measures. A preliminary study used 14 words generated by ChatGPT to gauge public attitudes toward AI in environmental decision-making, revealing a generally favorable view of AI in environmental management. Study 1a and Study 1b experimentally manipulated perceptions of AI’s unsustainability, demonstrating that awareness of its high energy consumption and carbon emissions significantly reduces acceptance. Study 2 replicated these findings and identified morality—rather than dependency—as the mediating factor. Study 3 showed that pro-environmental attitudes significantly moderate the relationship between AI sustainability and its acceptance in environmental contexts.

    In summary, this research highlights that while individuals are willing to collaborate with AI to address environmental challenges, their acceptance diminishes when AI is perceived as environmentally harmful. These findings underscore the critical importance of AI’s sustainability in achieving sustainable development goals.

    Figures and Tables | References | Related Articles | Metrics
    Emotional capabilities evaluation of multimodal large language model in dynamic social interaction scenarios
    ZHOU Zisen, HUANG Qi, TAN Zehong, LIU Rui, CAO Ziheng, MU Fangman, FAN Yachun, QIN Shaozheng
    2025, 57 (11):  1988-2000.  doi: 10.3724/SP.J.1041.2025.1988
    Abstract ( 1430 )   HTML ( 84 )  
    PDF (6426KB) ( 3641 )   Peer Review Comments

    Multimodal Large Language Models (MLLMs) can process and integrate multimodal data, such as images and text, providing a powerful tool for understanding human psychology and behavior. Combining classic emotional behavior experimental paradigms, this study compares the emotion recognition and prediction abilities of human participants and two mainstream MLLMs in dynamic social interaction contexts, aiming to disentangle the distinct roles of visual features of conversational characters (images) and conversational content (text) in emotion recognition and prediction.

    The results indicate that the emotion recognition and prediction performance of MLLMs, based on character images and conversational content, exhibits moderate or lower correlations with human participants. Despite a notable gap, MLLMs have begun to demonstrate preliminary capabilities in emotion recognition and prediction similar to human participants in dyadic interactions. Using human performance as a benchmark, the study further compares MLLMs under different conditions: integrating both character images and conversational content, using only character images, or relying solely on conversational content. The results suggest that visual features of character interactions somewhat constrain MLLMs’ basic emotion recognition but effectively facilitate the recognition of complex emotions, while having no significant impact on emotion prediction.

    Additionally, by comparing the emotion recognition and prediction performance of two mainstream MLLMs and different versions of GPT-4, the study finds that, rather than merely increasing the scale of training data, innovations in the underlying technical framework play a more crucial role in enhancing MLLMs’ emotional capabilities in dynamic social interaction contexts. Overall, this study deepens the understanding of the interaction between human visual features and conversational content, fosters interdisciplinary integration between psychology and artificial intelligence, and provides valuable theoretical and practical insights for developing explainable affective computing models and general artificial intelligence.

    Figures and Tables | References | Related Articles | Metrics
    The factors affecting subthreshold depression for people with occupational stress in the era of digital intelligence: Machine learning-based evidence
    DENG Lifang, PEI Bei, GAO Tian’ai
    2025, 57 (11):  2001-2021.  doi: 10.3724/SP.J.1041.2025.2001
    Abstract ( 2046 )   HTML ( 137 )  
    PDF (2243KB) ( 3134 )   Peer Review Comments

    Depression is one of the most common psychological problems, and subthreshold depression, as a precursor to its occurrence, plays a vital warning role in the prevention and treatment of depression. However, there is currently a lack of in-depth analysis of the representations and influencing factors of subthreshold depression in people with occupational stress in China (SDPOSC). This study integrated grounded-theory research with machine learning methods to explore the manifestations and influencing factors of subthreshold depression among Chinese working population under stress. BERT technology was utilized to construct a discriminant model for identifying the factors influencing subthreshold depression within this population, and the model's effectiveness was subsequently studied and confirmed.

    This research is composed of two studies. The first study involved the analysis of network texts harvested through web crawling, employing grounded theory for coding to establish a framework of factors influencing subthreshold depression in individuals under stress. The correlation structure between influencing factors and their representations was further explored, along with association rule analysis between influencing factors. Word frequency analysis and occupational difference tests were then conducted to analyze the characteristics of the influencing factors. Mann-Kendall test was subsequently applied to analyze the development trend of influencing factors. Based on the analysis of online text, the second study constructed a machine learning model using BERT technology. the influencing factors of subthreshold depression are distinguished and the effectiveness of the model is subsequently confirmed.

    Results showed that (1) Manifestations of subthreshold depression in people with occupational stress have five categories, with weakened willpower the highest frequency of expression and daily behavioral changes the lowest frequency. (2) Main influencing factors consist of eight categories, with work factors, evaluation adaptation, and autonomous selection the highest frequency; and stress events the lowest. (3) Eight types of influencing factors were closely related to subthreshold depression symptoms, with stressful event the best single predictor. Network analysis based on association rules revealed that "self-awareness, " "behavioral freedom, " "environmental adaptation, " and "general social interaction" are the most important subcategories. (4) Healthcare professionals had a significant difference in somatic factors compared to other professions, identified by a difference test of word frequency distribution. (5) Words related to work factors has shown an upward trend from 2011 to 2023, while those related to interpersonal factors have shown a downward trend. (6) A BERT-based machine learning model is obtained and it works in identifying influencing factors of subthreshold depression in populations experiencing work-related stress, in particular, the XGBoost algorithm achieved a prediction accuracy of 81.58%, with particularly strong performance in subthreshold depression detection (F1-score = 0.90, AUC = 0.93).

    This study provided an in-depth analysis of the representations and influencing factors of SDPOSC, enriching the localization research of subthreshold depression from an empirical perspective. Furthermore, a machine learning model by BERT can be utilized in subsequent research. The study of SDPOSC can help identify their depression risk and has important theoretical and practical significance for the prevention and treatment of SDPOSC.

    Figures and Tables | References | Related Articles | Metrics
    Self-help AI psychological counseling system based on large language models and its effectiveness evaluation
    HUANG Feng, DING Huimin, LI Sijia, HAN Nuo, DI Yazheng, LIU Xiaoqian, ZHAO Nan, LI Linyan, ZHU Tingshao
    2025, 57 (11):  2022-2042.  doi: 10.3724/SP.J.1041.2025.2022
    Abstract ( 2424 )   HTML ( 135 )  
    PDF (2804KB) ( 2781 )   Peer Review Comments

    The global prevalence of mental health issues, such as depression and anxiety, has become a significant public health challenge. Traditional mental health services face limitations in accessibility, affordability, and scalability. The emergence of large language models (LLMs) offers new opportunities for developing intelligent, self-help psychological counseling systems. However, optimizing LLMs for mental health applications presents unique challenges, including data scarcity and privacy concerns. This study aimed to address these challenges by constructing a self-help AI psychological counseling system using zero-shot learning and chain-of-thought prompting. It also evaluated the effectiveness of this established system in improving mental health outcomes among the general population. The research also explored the impact of AI anthropomorphizing on human- computer interaction outcomes in mental health interventions.

    The study comprised two parts. In Experiment 1, we constructed the AI counseling system based on the GPT-4o model. We first compared GPT-4o with two other LLMs (Claude 3 Opus and Yi-Large) using a test set of 12 common mental health topics covering interpersonal relationships, family issues, personal growth, and other categories. Three qualified psychological counselors evaluated the models' performance on normative quality, professionalism, emotional understanding and empathy, and consistency and coherence. We then optimized GPT-4o using chain-of-thought prompting and role instructions designed explicitly for psychological counseling scenarios. The optimized model was re-evaluated to assess improvements. In Experiment 2, we conducted a two-week randomized controlled trial with 202 participants from the general population who reported experiencing negative emotions or psychological distress but had not been diagnosed with severe mental health issues. Participants were randomly assigned to one of three experimental groups with varying degrees of AI anthropomorphizing (F: female counselor image and name, M: male counselor image and name, R: robot image without human name) or a control group (C: using unmodified GPT-4o). To ensure active participation, interactions with at least 10 dialogue rounds and spanning more than 10 minutes were considered valid for analysis. Mental health outcomes, including depression, anxiety, stress (measured by DASS-21), and loneliness (measured by SSL), were assessed at baseline (T1), the last two days of the one-week interaction (T2), and one week post-intervention (T3). Linear mixed-effects models were used to analyze the data, with simple effects analysis and Tukey HSD tests for post-hoc comparisons.

    In Experiment 1, GPT-4o significantly outperformed other models in normative quality, emotional understanding and empathy, and consistency and coherence (all p < 0.001). After optimization with chain-of-thought prompting, the model showed further significant improvements across all evaluation dimensions (p < 0.01), with huge effect sizes in normative quality (d = 1.28), emotional understanding and empathy (d = 1.06), and consistency and coherence (d = 1.14). Professional competence showed more limited improvement (d = 0.51), reflecting current technological limitations in this dimension. In Experiment 2, the attrition rate from T1 to T3 was 24.3%, with no significant differences in demographic characteristics or baseline mental health indicators between completers and non-completers. The interaction quality control retained 180 participants at T2(retention rate 89.11%) and 153 at T3(75.74%). All experimental groups showed significant short-term improvements in depression, anxiety, and loneliness at T2 compared to the control group (all p < 0.001). For loneliness, anthropomorphized AI designs (F and M groups) demonstrated significantly greater effects than the non-anthropomorphized design (R group) at T2. For stress levels, a group × time interaction effect reached marginally significant (p = 0.05), with only the non-anthropomorphized group (R group) showing substantial improvement from T1 to T2 (b = 2.35, SE = 0.48, p < 0.001). The improvement in anxiety symptoms persisted at T3 for all experimental groups (p < 0.001), while effects on depression, stress, and loneliness did not maintain significance at follow-up.

    This study provides empirical evidence for the potential of AI-based self-help psychological counseling in improving mental health outcomes, particularly in reducing mental health symptoms in the short term. The successful application of zero-shot learning and chain-of-thought prompting in optimizing LLMs for mental health dialogues offers a novel approach to overcome challenges in data scarcity and model adaptation in specialized domains. The differential effects of AI anthropomorphization on various mental health indicators support a nuanced design framework: anthropomorphized designs may be more effective for addressing social functioning-related issues like loneliness through enhanced social presence. In contrast, non-anthropomorphized designs might better manage stress by reducing social evaluation pressure. However, the study also reveals significant limitations, including the lack of long-term effects for most outcomes and limited improvement in professional competence. Future research should focus on enhancing the long-term efficacy of AI-assisted mental health interventions, improving professional depth for specialized counseling scenarios, exploring human-AI collaborative models for high-risk cases, and further investigating the mechanisms underlying the differential effects of AI design features on specific mental health issues. These findings provide valuable insights for developing more effective, personalized AI-assisted mental health services to complement traditional care approaches.

    Figures and Tables | References | Related Articles | Metrics
    Humans perceive warmth and competence in large language models
    WU Yueting, WANG Bo, BAO Han Wu Shuang, LI Ruonan, WU Yi, WANG Jiaqi, CHENG Cheng, YANG Li
    2025, 57 (11):  2043-2059.  doi: 10.3724/SP.J.1041.2025.2043
    Abstract ( 1683 )   HTML ( 73 )  
    PDF (579KB) ( 2191 )   Peer Review Comments

    The rapid development and application of Large Language Models (LLMs) have significantly enhanced their capabilities, influencing human-machine interactions in profound ways. As LLMs evolve, society is shifting from traditional interpersonal interactions to a multilayered structure integrating human-to-human, human-to-machine, and machine-to-machine interactions. In this context, understanding how humans perceive and evaluate LLMs—and whether this follows the Big Two model of warmth and competence in interpersonal perception—has become critical. This study examines human perceptions of LLMs through three progressive empirical studies.

    Participants with prior LLM experience were recruited for the studies. Study 1 comprised two sub-studies: Study 1a (N = 207) used a free-response task, asking participants to describe their impressions of LLMs using at least three words, which were analyzed using the Semi-Automated Dictionary Creation for Analyzing Text to identify key dimensions of perception. Study 1b (N = 219) involved a lexical rating task, in which participants rated the applicability of selected evaluation words to LLMs. Study 2 (N = 178) used a questionnaire, in which participants rated a familiar LLM and provided feedback on their willingness to continue using it and their liking of it. Study 3 (N = 207) employed a questionnaire survey to assess participants’ ratings of warmth and competence for both humans and LLMs.

    Study 1 found that humans primarily perceive LLMs through warmth and competence, similar to how they perceive other humans. In general contexts, participants prioritized competence over warmth when evaluating LLMs, showing a significant priority effect (odds ratio = 2.88, z = 9.512, 95% CI [2.32, 3.59], p< 0.001). This contrasts with the typical warmth-priority effect in human-to-human perception. Study 2 investigated the relationship between perceptions of warmth and competence and human attitudes toward LLMs, specifically their emotional (e.g., liking) and functional (e.g., willingness to continue using) attitudes. Results showed that both dimensions positively predicted participants’ liking and willingness to continue using LLMs. Warmth had a stronger predictive effect on liking (warmth: β = 0.41, p< 0.001; competence: β = 0.27, p< 0.001), while competence had a stronger predictive effect on willingness to continue using (warmth: β = 0.19, p= 0.005; competence: β = 0.45, p< 0.001). This outcome suggests that the priority effect of warmth and competence shifts across attitude predictions. Study 3 examined specific LLMs ratings in terms of warmth and competence. Results showed no significant difference in warmth ratings between humans (M= 5.06, SD= 1.09) and LLMs (M= 5.11, SD= 1.23), t(206) = −0.60, p= 0.551. However, LLMs were rated significantly higher on competence (M= 5.16, SD= 1.20) than humans (M= 4.81, SD= 1.23), t(206) = −3.51, p< 0.001, Cohen’s d = −0.29.

    This study makes two significant contributions to the field. First, it establishes a preliminary theoretical framework for understanding human perception of LLMs. Second, it offers new insights into human-machine interaction by emphasizing the importance of warmth and competence in shaping user attitudes. The findings have practical implications for AI design and policymaking, providing a framework for improving user acceptance, optimizing LLM design, and promoting responsible human-AI coexistence.

    Figures and Tables | References | Related Articles | Metrics
    Employees adhere less to advice on moral behavior from artificial intelligence supervisors than human
    XU Liying, ZHAO Yijun, YU Feng
    2025, 57 (11):  2060-2082.  doi: 10.3724/SP.J.1041.2025.2060
    Abstract ( 1313 )   HTML ( 74 )  
    PDF (672KB) ( 1556 )   Peer Review Comments

    The use of artificial intelligence (AI) in organizations has evolved from being a tool to being a supervisor. Although previous research has examined people's reactions to AI supervisors in general, few studies have investigated the effectiveness of AI supervisors, specifically whether individuals adhere to their moral behavioral advices. The present research aims to compare employees' adherence to moral behavioral advice given by AI and human supervisors, as well as identify potential psychological mechanisms and boundary conditions behind the possible differences.

    To test our research hypotheses, we conducted six experiments and three pilot experiemts (N = 1642, including 179 samples of pilot experiments) involving different types of moral behaviors in organizations, such as engaging in the activity to help the disabled, volunteering for environmental protection or child welfare, and making charitable donations for disasters or colleagues' difficulties. Experiment 1a and 1b was a single-factor, two-level, between-subjects design. 180 participants were randomly assigned to two conditions: supervisors giving advice on moral behavior (human versus AI). Their adherence to the supervisor's advice was measured in different scenarios. Experiment 2 followed the same design as Experiment 1, with additional measurements of evaluation apprehension and perceived mind to test the mediating role. To establish a causal chain between the mediator and the dependent variable and demonstrate the robustness of our findings, we further examined the underlying mechanism in Experiment 3. This experiment had a between-subjects design of 2 (supervisors: human versus AI) × 2 (evaluation apprehension: high versus low). Experiments 4 and 5 were designed to test the moderating role of anthropomorphism. In Experiment 4, participants' tendency to anthropomorphize was measured, and in Experiment 5, the anthropomorphism of the AI supervisor was manipulated.

    As predicted, the present research found that, compared to a human supervisor, participants were less likely to follow the moral advice of an AI supervisor (Experiments 1a~5). The robustness of this finding was demonstrated by the diversity of our scenario settings and samples. And we also excluded the potential effects of perceived rational, negative emotions, exploitation, perceived autonomy and some individual differences (pilot experiment and emperiment 1a~1b). In addition, this research discovered evaluation apprehension as the underlying mechanism explaining employees' adherence to advice from different supervisors. Participants believed that they would receive less social judgment and evaluation from an AI supervisor than a human supervisor. Consequently, they were less willing to adhere to the advice offered by the AI (Experiments 2~5). The present research also demonstrated the moderating effect of anthropomorphism (Experiments 4~5). In Experiment 4, for individuals with a high tendency towards anthropomorphism, there was no significant difference in their adherence to advice on moral behavior from human or AI supervisors; Participants with low anthropomorphism tendency showed greater adherence to a human supervisor than to an AI supervisor. In Experiment 5, participants demonstrated greater adherence to the AI supervisor with a human-like name and communication style compared to the mechanized AI supervisor.

    The study contributes to the literature on AI leadership by highlighting the limitations of AI supervisors in providing advice on moral behavior. Additionally, the results confirm the phenomenon of algorithm aversion in the moral domain, indicating that people are hesitant to accept AI involvement in moral decision-making, even in an advisory role. The study also identifies evaluation apprehension as a factor that influences adherence to AI advice. Individuals may be less likely to follow the advice of AI due to a decreased concern for potential social judgment in their interactions with AI supervisors. Finally, anthropomorphism may be a useful approach to enhance the effectiveness of AI supervisors.

    Figures and Tables | References | Related Articles | Metrics