班级(哲学)
选择(遗传算法)
计算机科学
人工智能
心理学
作者
Walter Laurito,Ben Davis,Peli Grietzer,Tomáš Gavenčiak,Alina Böhm,Jan Kulveit
标识
DOI:10.1073/pnas.2415697122
摘要
Are large language models (LLMs) biased in favor of communications produced by LLMs, leading to possible antihuman discrimination? Using a classical experimental design inspired by employment discrimination studies, we tested widely used LLMs, including GPT-3.5, GPT-4 and a selection of recent open-weight models in binary choice scenarios. These involved LLM-based assistants selecting between goods (the goods we study include consumer products, academic papers, and film-viewings) described either by humans or LLMs. Our results show a consistent tendency for LLM-based AIs to prefer LLM-presented options. This suggests the possibility of future AI systems implicitly discriminating against humans as a class, giving AI agents and AI-assisted humans an unfair advantage.
科研通智能强力驱动
Strongly Powered by AbleSci AI