採用におけるAIバイアスが人間の意思決定にも反映される可能性を発見(People mirror AI systems’ hiring biases, study finds)

2025-11-10 ワシントン大学(UW)

UWの研究チームは、528人の参加者を対象に、16職種(コンピュータシステムアナリスト、ナースプラクティショナー、ハウスキーパーなど)でAI(模擬大規模言語モデル)と協働して候補者を選ぶ実験を行い、AIの推奨に人間の意思決定がどう影響されるかを調査した。研究では、同じ能力を持つ白人男性・黒人男性・ヒスパニック男性・アジア系男性の履歴書を用い、AIから「人種的に中立」「やや偏りあり」「強い偏り」の3条件の推薦を提示。AI推薦が中立の場合、参加者の選択には偏りが見られなかったが、AIが特定人種を推薦すると、参加者もその偏りに沿った選択を行った。特に強い偏りの条件では、参加者はAI推薦とほぼ同様の偏向判断を下していた。研究者は、「多くの企業が人によるレビューなしにAIの推薦を用いており、人−AIの協働モデルが支配的になっている。だが人間はAIの偏りを自ら修正することなく、提示された偏りをそのまま受け入れやすい」と警告している。

<関連情報>

思考なし、AIのみ:偏ったLLM採用推奨は人間の意思決定を変え、人間の自律性を制限する No Thoughts Just AI: Biased LLM Hiring Recommendations Alter Human Decision Making and Limit Human Autonomy

Kyra Wilson,Mattea Sim,Anna-Maria Gueorguieva,Aylin Caliskan
Proceedings of the Eighth AAAI/ACM Conference on AI, Ethics, and Society (AIES-25)
DOI:https://doi.org/10.1609/aies.v8i3.36749

Abstract

Despite bias in artificial intelligence (AI) being a risk of their use in hiring systems, there is no large-scale empirical investigation of the impacts of these biases on hiring decisions made collaboratively between people and AI systems. It is also unknown whether AI literacy, people’s own biases, and behavioral interventions intended to reduce discrimination affect these human-in-the-loop AI teaming (AI-HITL) outcomes. In this study, we conduct a resume-screening experiment (N=528) where people collaborate with simulated AI models exhibiting race-based preferences (bias) to evaluate candidates for 16 high and low status occupations. Simulated AI bias approximates factual and counterfactual estimates of racial bias in real-world AI systems. We investigate people’s preferences for White, Black, Hispanic, and Asian candidates (represented through names and affinity groups on quality-controlled resumes) across 1,526 scenarios and measure their unconscious associations between race and status using implicit association tests (IATs), which predict discriminatory hiring decisions but have not been investigated in human-AI collaboration. This evaluation framework can generalize to other groups, models, and domains. When making decisions without AI or with AI that exhibits no race-based preferences, people select all candidates at equal rates. However, when interacting with AI favoring a particular group, people also favor those candidates up to 90\% of the time, indicating a significant behavioral shift. The likelihood of selecting candidates whose identities do not align with common race-status stereotypes can increase by 13\% if people complete an IAT before conducting resume screening. Finally, even if people think AI recommendations are low quality or not important, their decisions are still vulnerable to AI bias under certain circumstances. This work has implications for people’s autonomy in AI-HITL scenarios, AI and work, design and evaluation of AI hiring systems, and strategies for mitigating bias in collaborative decision-making tasks. In particular, organizational and regulatory policy should acknowledge the complex nature of AI-HITL decision making when implementing these systems, educating people who use them, and determining which are subject to oversight.

1603情報システム・データ工学
ad
ad
Follow
ad
タイトルとURLをコピーしました