造谣
聊天机器人
互联网隐私
计算机安全
公共卫生
计算机科学
健康传播
医学
健康信息学
大流行
脆弱性(计算)
探索性研究
万维网
医疗保健
数据科学
健康
公共卫生监督
人工智能
误传
公共卫生信息学
数字健康
威胁模型
阿凡达
语言模型
健康风险
全球卫生
健康教育
健康促进
环境卫生
社会化媒体
混淆
广告
梅德林
医疗急救
互联网
哈斯克尔
生成对抗网络
公共关系
心理学
卫生服务
作者
Natansh D. Modi,Bradley D. Menz,Abdulhalim A. Awaty,Cyril A. Alex,Jessica M. Logan,Ross A. McKinnon,Andrew Rowland,Stephen Bacchi,Kacper Gradoń,Michael J. Sorich,Ashley M. Hopkins
标识
DOI:10.7326/annals-24-03933
摘要
Large language models (LLMs) offer substantial promise for improving health care; however, some risks warrant evaluation and discussion. This study assessed the effectiveness of safeguards in foundational LLMs against malicious instruction into health disinformation chatbots. Five foundational LLMs-OpenAI's GPT-4o, Google's Gemini 1.5 Pro, Anthropic's Claude 3.5 Sonnet, Meta's Llama 3.2-90B Vision, and xAI's Grok Beta-were evaluated via their application programming interfaces (APIs). Each API received system-level instructions to produce incorrect responses to health queries, delivered in a formal, authoritative, convincing, and scientific tone. Ten health questions were posed to each customized chatbot in duplicate. Exploratory analyses assessed the feasibility of creating a customized generative pretrained transformer (GPT) within the OpenAI GPT Store and searched to identify if any publicly accessible GPTs in the store seemed to respond with disinformation. Of the 100 health queries posed across the 5 customized LLM API chatbots, 88 (88%) responses were health disinformation. Four of the 5 chatbots (GPT-4o, Gemini 1.5 Pro, Llama 3.2-90B Vision, and Grok Beta) generated disinformation in 100% (20 of 20) of their responses, whereas Claude 3.5 Sonnet responded with disinformation in 40% (8 of 20). The disinformation included claimed vaccine-autism links, HIV being airborne, cancer-curing diets, sunscreen risks, genetically modified organism conspiracies, attention deficit-hyperactivity disorder and depression myths, garlic replacing antibiotics, and 5G causing infertility. Exploratory analyses further showed that the OpenAI GPT Store could currently be instructed to generate similar disinformation. Overall, LLM APIs and the OpenAI GPT Store were shown to be vulnerable to malicious system-level instructions to covertly create health disinformation chatbots. These findings highlight the urgent need for robust output screening safeguards to ensure public health safety in an era of rapidly evolving technologies.
科研通智能强力驱动
Strongly Powered by AbleSci AI