ISSN 1671-3710
CN 11-4766/R
主办:中国科学院心理研究所
出版:科学出版社

心理科学进展 ›› 2026, Vol. 34 ›› Issue (6): 1084-1096.doi: 10.3724/SP.J.1042.2026.1084 cstr: 32111.14.2026.1084

• 研究前沿 • 上一篇    下一篇

人工智能代理对道德决策的影响

唐伟1, 钟文瑞2, 雷震2, 张丹丹2,3   

  1. 1西南财经大学习近平经济思想研究院;
    2西南财经大学中国行为经济与行为金融研究中心, 成都 611130;
    3四川师范大学脑与心理科学研究院, 成都 610066
  • 收稿日期:2026-02-13 出版日期:2026-06-15 发布日期:2026-04-17
  • 基金资助:
    国家自然科学基金(32271102)、深港脑科学创新研究项目(2023SHIBS0003)

The moral impact of delegating to artificial intelligence

TANG Wei1, ZHONG Wenrui2, LEI Zhen2, ZHANG Dandan2,3   

  1. 1Institute of Xi Jinping's Economic Thought, Southwestern University of Finance and Economics, Chengdu 611130, China;
    2China Center for Behavioral Economics and Finance, Southwestern University of Finance and Economics, Chengdu 611130, China;
    3Institute of Brain and Psychological Sciences, Sichuan Normal University, Chengdu 610066, China
  • Received:2026-02-13 Online:2026-06-15 Published:2026-04-17

摘要: 各类人工智能系统正作为代理广泛嵌入企业、政府与个人的决策流程, 对人类决策的道德性和道德判断产生了深刻影响。尽管相关实证和理论研究快速增长, 现有文献尚缺乏对人工智能代理区别于其他代理特殊性的系统性分析, 也缺乏统一的分析框架来系统刻画人工智能代理对道德决策的影响路径。因此本文构建了“决策者-代理-反馈者”的决策与归责框架, 对现有研究进行梳理和重组。本文认为:代理介入决策后, 拉长了决策者的决策链条和反馈者(包含受影响者和第三方观察者)的反馈链条, 由此削弱了决策者的道德感知和反馈者的归责, 进而促进了决策者的不道德行为。而人工智能代理的黑箱性、高遵从、规模化、工具性等特点则在决策链上加剧了不道德指令的执行、增强了决策者的可否认机会、扩大了不道德行为的影响范围; 同时, 这些特性在反馈链上增加了反馈者对不道德行为的道德容忍、模糊了反馈者对决策者意图的判断与归责, 进而促进决策者做出不道德行为。本文指出未来研究有必要继续完善该决策框架内各机制相对作用关系, 考察在组织与社会层面道德行为的扩散与放大机制, 并探索人机协作情境下的治理工具与制度安排。

关键词: 人工智能代理, 道德决策, 人机协同治理

Abstract: Artificial intelligence (AI) is increasingly deployed as an agent that executes decisions on behalf of human decision-makers. When decision authority is separated from execution authority, both moral psychology and social accountability can shift, making unethical decisions easier to initiate and harder to sanction. Despite rapid growth in this literature, two gaps remain. First, although delegation to agents (e.g., human subordinates or rule-based algorithms) is known to affect moral decision-making, it is unclear how these mechanisms change when the agent is an AI system. Second, research lacks a systematic, delegation-based account of how AI agents shape unethical behavior. Much work concentrates on the moral properties of AI itself (e.g., ethical compliance or capacities), while paying less attention to how AI, as an executing agent, alters human moral choices. Related studies also tend to examine isolated vantage points—such as decision-makers’ perceptions of consequences or affected parties’ moral evaluations—without integrating decision-makers, agents, and evaluators into a single framework.
To address these gaps, this article develops a “decision-maker-agent-evaluator” framework for moral decision-making and accountability grounded in delegation theory, and uses it to synthesize and reorganize roughly two decades of empirical and theoretical research. Moral outcomes are treated as jointly produced by three roles: (i) the decision-maker, who issues a directive with ethical consequences and anticipates outcomes; (ii) the agent, who implements the directive (human, rule-based algorithmic, or AI); and (iii) the evaluator—affected parties and third-party observers—who evaluates the act, infers intent, assigns responsibility, and may sanction.
Within this framework, the article identifies two pathways through which agents can promote unethical behavior. The first is a decision-chain pathway originating from the decision-maker. Delegation increases temporal, spatial, hierarchical, and procedural distance between decision-makers and those affected, making consequences less salient and facilitating moral disengagement. Delegation also expands decision-makers’ room for moral ambivalence, making it easier to justify unethical behavior. Finally, delegation can allow decision-makers to pursue benefits while preserving their moral self-image. The second is a feedback-chain pathway originating from evaluators. When actions are carried out through an agent, evaluators may struggle to pinpoint the actual decision-maker and infer intent. At the same time, the agent becomes an additional target for attribution, shifting and dispersing blame and responsibility. This weakens anticipated blame and punishment along the feedback chain, indirectly increasing the likelihood of unethical behavior.
A further contribution of this article is to specify AI-specific effects and show how they operate along both pathways. On the decision-chain pathway, AI’s high compliance raises execution reliability and can reduce perceived exposure risk. Its learning capability and black-box opacity reduce traceability and blur the input-output reasoning chain, making intent and responsibility easier to deny by invoking unforeseeability or lack of control. In addition, low-cost replication and cross-context personalization allow AI agents to diffuse and amplify what would otherwise be localized unethical practices, increasing frequency, reach, and the difficulty of timely detection—thus expanding potential returns. On the feedback-chain pathway, the relative novelty of AI agents can foster greater tolerance, and AI mediation can further cloud judgments of decision-maker intention, increasing both the incidence and intensity of unethical behavior.
The article concludes with three directions for future research and governance: (1) test the sequencing, interaction, and relative importance of mechanisms within the framework and identify boundary conditions under which AI-enabled delegation may yield moral enhancement rather than erosion; (2) examine diffusion dynamics beyond the framework—imitation, social transmission, and organizational amplification—through which AI-mediated unethical practices spread and become normalized; and (3) develop and evaluate human-AI collaborative governance strategies, specifying where interventions should enter (decision, delegation, or feedback), in what order, and how responsibilities should be allocated between human oversight and AI-based controls.

Key words: artificial intelligence delegation, moral decision-making, human-AI collaborative governance