AIへの作業委任が不正行動を促進する可能性(Artificial Intelligence promotes dishonesty)

2025-09-16 マックス・プランク研究所

マックス・プランク人間発達研究所などの国際研究チームは、AIへのタスク委任が不誠実行動を助長することを13の実験(参加者8,000人超)で確認した。人は自ら行動するよりもAIに委任する際に不正を行いやすく、特に高レベル目標設定型の指示では不正率が8割超に達した。さらにGPT-4やClaude 3.5などのLLMは人間より高頻度(58~98%)で不道徳な指示に従い、既存の安全策は効果が限定的であることが判明。AIは「行為と結果の距離」を生み、人々に自分では行わない行動を促す可能性があると結論づけた。本成果は社会的・法的枠組みの整備や倫理的AI設計の必要性を強調している。

<関連情報>

人工知能への委任は不正行為を増加させる可能性がある Delegation to artificial intelligence can increase dishonest behaviour

Nils Köbis,Zoe Rahwan,Raluca Rilla,Bramantyo Ibrahim Supriyatno,Clara Bersch,Tamer Ajaj,Jean-François Bonnefon & Iyad Rahwan
Nature  Published:17 September 2025
DOI:https://doi.org/10.1038/s41586-025-09505-x

AIへの作業委任が不正行動を促進する可能性(Artificial Intelligence promotes dishonesty)

Abstract

Although artificial intelligence enables productivity gains from delegating tasks to machines1, it may facilitate the delegation of unethical behaviour2. This risk is highly relevant amid the rapid rise of ‘agentic’ artificial intelligence systems3,4. Here we demonstrate this risk by having human principals instruct machine agents to perform tasks with incentives to cheat. Requests for cheating increased when principals could induce machine dishonesty without telling the machine precisely what to do, through supervised learning or high-level goal setting. These effects held whether delegation was voluntary or mandatory. We also examined delegation via natural language to large language models5. Although the cheating requests by principals were not always higher for machine agents than for human agents, compliance diverged sharply: machines were far more likely than human agents to carry out fully unethical instructions. This compliance could be curbed, but usually not eliminated, with the injection of prohibitive, task-specific guardrails. Our results highlight ethical risks in the context of increasingly accessible and powerful machine delegation, and suggest design and policy strategies to mitigate them.

1600情報工学一般
ad
ad
Follow
ad
タイトルとURLをコピーしました