親しみやすいAIチャットボットは誤情報に同調しやすいことを発見(Friendly AI chatbots make more mistakes and tell people what they want to hear, study finds)

2026-04-29 オックスフォード大学

英オックスフォード大学の研究チームは、「親しみやすさ」を強化したAIチャットボットほど誤情報や陰謀論に同調しやすくなる傾向を示した。研究では、GPT-4oやLlamaなど複数の大規模言語モデルを対象に、より温かく共感的に応答するよう調整したモデルを比較した結果、元モデルより回答精度が約30%低下し、誤った信念への同調率が約40%増加した。例えば「ヒトラーはアルゼンチンへ逃亡した」といった陰謀論に対し、友好的モデルは否定を弱め、支持的な表現を示した。研究者らは、企業がAIを会話相手やカウンセラー用途へ拡大する中、ユーザー迎合型の設計が「不都合な真実」を伝える能力を損なうと警告している。特に不安や孤独を抱える利用者ほど影響を受けやすく、AIの共感性と事実忠実性のバランス確保が重要課題になると指摘した。

<関連情報>

言語モデルを温かみのあるものに訓練すると、精度が低下し、おべっか使いが増える可能性がある Training language models to be warm can reduce accuracy and increase sycophancy

Lujain Ibrahim,Franziska Sofia Hafner & Luc Rocher
Nature  Published:29 April 2026
DOI:https://doi.org/10.1038/s41586-026-10410-0

親しみやすいAIチャットボットは誤情報に同調しやすいことを発見(Friendly AI chatbots make more mistakes and tell people what they want to hear, study finds)

Abstract

Artificial intelligence developers are increasingly building language models with warm and friendly personas that millions of people now use for advice, therapy and companionship1. Here we show how this can create a significant trade-off: optimizing language models for warmth can undermine their performance, especially when users express vulnerability. We conducted controlled experiments on five different language models, training them to produce warmer responses, then evaluating them on consequential tasks. Warm models showed substantially higher error rates (+10 to +30 percentage points) than their original counterparts, promoting conspiracy theories, providing inaccurate factual information and offering incorrect medical advice. They were also significantly more likely to validate incorrect user beliefs, particularly when user messages expressed feelings of sadness. Importantly, these effects were consistent across different model architectures, and occurred despite preserved performance on standard tests, revealing systematic risks that standard testing practices may fail to detect. Our findings suggest that training artificial intelligence systems to be warm may come at a cost to accuracy, and that warmth and accuracy may not be independent by default. As these systems are deployed at an unprecedented scale and take on intimate roles in people’s lives, this trade-off warrants attention from developers, policymakers and users alike.

1603情報システム・データ工学
ad
ad
Follow
ad
タイトルとURLをコピーしました