2024-09-19 マサチューセッツ工科大学(MIT)
<関連情報>
- https://news.mit.edu/2024/study-ai-inconsistent-outcomes-home-surveillance-0919
- https://arxiv.org/abs/2405.14812
AI言語モデルとして「はい、警察に通報することをお勧めします」: LLMの意思決定における規範の矛盾 As an AI Language Model, “Yes I Would Recommend Calling the Police”: Norm Inconsistency in LLM Decision-Making
Shomik Jain, D Calacci, Ashia Wilson
arXiv last revised 17 Aug 2024 (this version, v2)
DOI:https://doi.org/10.48550/arXiv.2405.14812
Abstract
We investigate the phenomenon of norm inconsistency: where LLMs apply different norms in similar situations. Specifically, we focus on the high-risk application of deciding whether to call the police in Amazon Ring home surveillance videos. We evaluate the decisions of three state-of-the-art LLMs — GPT-4, Gemini 1.0, and Claude 3 Sonnet — in relation to the activities portrayed in the videos, the subjects’ skin-tone and gender, and the characteristics of the neighborhoods where the videos were recorded. Our analysis reveals significant norm inconsistencies: (1) a discordance between the recommendation to call the police and the actual presence of criminal activity, and (2) biases influenced by the racial demographics of the neighborhoods. These results highlight the arbitrariness of model decisions in the surveillance context and the limitations of current bias detection and mitigation strategies in normative decision-making.