OpenAI’s ChatGPT and Its Role in Plastic Surgery Research

医学 整形外科 普通外科 外科
作者
Allan A. Weidman,Lauren Valentine,Kevin C. Chung,Samuel J. Lin
出处
期刊:Plastic and Reconstructive Surgery [Ovid Technologies (Wolters Kluwer)]
卷期号:151 (5): 1111-1113
标识
DOI:10.1097/prs.0000000000010342
摘要

Artificial intelligence (AI), which once existed solely in science fiction, has arrived as a part of our daily lives. In the past year, various AI models have been released for public use, with sweeping ramifications. One company in particular, OpenAI, has offered two popular AI models, DALL-E and ChatGPT, to the public for free. DALL-E is an artificial intelligence model that produces original computer-generated images, while ChatGPT generates text while interacting with users in a conversational way.1 For example, a user might say, “Hi ChatGPT,” and ChatGPT might write back, “Hello, how are you today?” Or a user could say, “ChatGPT, write an essay on deep inferior epigastric perforator (DIEP) flaps,” and ChatGPT would respond within seconds with a well-written, knowledgeable-sounding essay about DIEP flaps. Interestingly, ChatGPT was not simply created all at once but needed to be trained using what OpenAI calls “reinforcement learning from human feedback.”1 During this training, ChatGPT was exposed to a wide array of human text and information to give it the base of knowledge needed to write such an essay about DIEP flaps. Given these wide-ranging capabilities, the model has quickly gained massive popularity. After being launched on November 30, 2022, the research release of ChatGPT reached 1 million users in just 5 days.2 In comparison, Facebook took approximately 10 months to reach 1 million users.3 Although the model is currently free to the public, the service’s founder has already stated that eventually efforts will be made to generate revenue from the model. In addition, OpenAI recently strengthened its partnership with Microsoft after receiving a multibillion-dollar investment and will likely continue to expand.4 Innovative uses of ChatGPT have been documented across a variety of fields, including medicine. In one example, a physician shared a TikTok video of himself using the AI tool to quickly generate prior authorization letters, albeit with significant editing required.5 In another scenario, a mental health app used ChatGPT to help provide mental health support to approximately 4000 people.5 This experiment drew significant backlash, given the ethical implications of the study, but highlights a larger trend in which people report turning to AI as a source of therapy.6,7 Furthermore, ChatGPT was able to perform at a level comparable to a third-year medical student on National Board of Medical Examiners examinations and pass the United States Medical Licensing Examination Step examinations.8,9 Finally, in a much simpler example and as a demonstration of how ChatGPT functions, we asked the model to write a poem about the Plastic and Reconstructive Surgery journal using the command, “Write a poem about the Plastic and Reconstructive Surgery journal” (Fig. 1). Having drawn on its knowledge of poem conventions and the journal gained during its training, either inferred or explicitly taught, ChatGPT wrote a new poem that is completely original. It did not need to be trained on the topic of plastic and reconstructive surgery, nor did it require a program that is specifically designed to write poetry. If one were to ask ChatGPT to write the poem again using the same prompt, it would generate a different original poem altogether.Fig. 1.: Poem generated by ChatGPT.As with all emerging technology, ChatGPT has significant limitations. Primarily, the AI model can generate incorrect and misleading information while appearing confident about the veracity of the material. As a result, if not fact checked and edited rigorously, fallacious material can be unknowingly distributed by ChatGPT users. Accordingly, OpenAI’s website lists limitations for ChatGPT, the first of which is, “ChatGPT sometimes writes plausible-sounding but incorrect or nonsensical answers.”1 This was recently highlighted in a case report that asked 15 radiologists to assess ChatGPT’s ability to simplify radiology reports. While most reports were factually correct and complete, about half of the radiologists found statements in the generated reports with a “high potential of leading patients to the wrong conclusion” and various instances of errors that led to the exclusion of key medical findings.10 For example, ChatGPT often interpreted “differential diagnosis” as the final diagnosis, falsely leading patients to believe that they had certain conditions that were merely one of many possibilities. Another report falsely stated that there was no evidence of cancer spreading to other parts of the body, when evidence of pulmonary metastases was documented in the original report.10 Further, as it is not freely connected to the internet, ChatGPT’s knowledge base is limited to the information it was trained on (exposed to) during its creation; therefore, it is ignorant of current events and new advances.11 Similarly, as the model only draws knowledge from the human-generated information it was trained on, any biases present in that information may also be present in ChatGPT’s outputs. Thus, it does not necessarily deliver the right answer but rather answers based on the information that it acquired and inferred from its training. A simplified example is that if ChatGPT were exposed to a racist blog during its training, it may then generate racist dialogue when asked certain questions. Caliskan et al.12 highlighted this principle by using the Implicit Association Test to replicate a series of known human biases.13 The authors then demonstrated that machines can learn implicit associations that are exhibited by humans, and thus have the risk of perpetuating dangerous biases and stereotypes.12 OpenAI has taken efforts to address the issue of bias, but has acknowledged that it remains a problem, stating, “While we’ve made efforts to make the model refuse inappropriate requests, it will sometimes respond to harmful instructions or exhibit biased behavior.”1 The consequences of incorrect and biased material generated by AI platforms led to significant hardships for another AI model called Galactica, which was created by Facebook’s parent company, Meta. Galactica is a language model similar to ChatGPT that was trained using scientific literature. Following its public release in mid-November of 2022, however, it was quickly removed from public access days later after it faced criticism for presenting biased and incorrect information as fact.14 As Meta works to reform the model, Galactica serves as an ongoing reminder of the limitations of other AI tools similar to ChatGPT. Given its utility and despite its flaws, ChatGPT has been used in scientific research and has subsequently become a hotly debated topic among peer-reviewed journals. ChatGPT has been applied to prepare manuscripts, and in several cases, it was even listed as an author.15 As a result, one study investigated whether medical researchers (ie, potential editors) could identify AI-generated abstracts from human-generated abstracts. The study found human reviewers believed incorrectly that 32% of AI-generated texts had been written by humans.16 Thus, medicine and science stand at an impasse, with the ethics of utilizing AI models such as ChatGPT to write scientific text still not clearly defined. However, multiple publishers and journals have agreed that ChatGPT and other AI models do not meet the legal requirements of an author.15 Technology is wonderful as a tool. There are aspects of human communication that should be left to human beings. We will need to see over time how natural language processing is or is not incorporated into life. The exercise of writing causes writers to deeply examine and to reflect on their data analysis and their message globally. Plastic and Reconstructive Surgery holds the position that the use of ChatGPT and other AI models should not be encouraged, nor should they be included as an author on scientific articles. Authors must use extreme caution and judicious ethical discretion when engaging ChatGPT or other AI services to generate academic writing or other portions of their study. Authors are personally liable for any mistakes made by the tool but reported by them as fact. If AI tools are used, authors should disclose their use in the acknowledgments on the title page, as opposed to delegating authorship. Finally, investigators must work closely with their institutional review boards to ensure that the use of AI in their research is ethical, especially with regard to security of patient information. DISCLOSURE The authors have no financial interest to declare in relation to the content of this article.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
更新
大幅提高文件上传限制,最高150M (2024-4-1)

科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
1秒前
舒心的怡完成签到,获得积分20
1秒前
wbb发布了新的文献求助10
1秒前
Georgechan发布了新的文献求助10
3秒前
3秒前
SciGPT应助liu采纳,获得10
3秒前
梁吃鱼完成签到,获得积分10
4秒前
kkk发布了新的文献求助10
5秒前
cuicui发布了新的文献求助10
5秒前
6秒前
7秒前
9秒前
芈钥完成签到 ,获得积分10
9秒前
KEyanba完成签到,获得积分10
10秒前
10秒前
chengmin发布了新的文献求助10
11秒前
养乐多完成签到,获得积分10
12秒前
13秒前
13秒前
虾滑完成签到,获得积分10
13秒前
隐形曼青应助科研通管家采纳,获得10
15秒前
科研通AI2S应助科研通管家采纳,获得10
15秒前
乐乐应助科研通管家采纳,获得10
15秒前
科目三应助科研通管家采纳,获得10
15秒前
紫金大萝卜应助JIEUN采纳,获得20
15秒前
我是老大应助科研通管家采纳,获得10
15秒前
爆米花应助科研通管家采纳,获得30
15秒前
orixero应助科研通管家采纳,获得10
16秒前
17秒前
爆米花应助莫寻雪采纳,获得10
17秒前
俊逸芸遥发布了新的文献求助10
18秒前
小可爱发布了新的文献求助10
19秒前
labulabu应助GZX采纳,获得10
20秒前
欣喜莫茗完成签到 ,获得积分10
20秒前
20秒前
20秒前
研友Zby14n完成签到 ,获得积分10
20秒前
21秒前
chen发布了新的文献求助10
21秒前
香蕉觅云应助单纯的沛白采纳,获得10
23秒前
高分求助中
【本贴是提醒信息,请勿应助】请在求助之前详细阅读求助说明!!!! 20000
One Man Talking: Selected Essays of Shao Xunmei, 1929–1939 1000
The Three Stars Each: The Astrolabes and Related Texts 900
Yuwu Song, Biographical Dictionary of the People's Republic of China 800
Multifunctional Agriculture, A New Paradigm for European Agriculture and Rural Development 600
Challenges, Strategies, and Resiliency in Disaster and Risk Management 500
Bernd Ziesemer - Maos deutscher Topagent: Wie China die Bundesrepublik eroberte 500
热门求助领域 (近24小时)
化学 材料科学 医学 生物 有机化学 工程类 生物化学 纳米技术 物理 内科学 计算机科学 化学工程 复合材料 遗传学 基因 物理化学 催化作用 电极 光电子学 量子力学
热门帖子
关注 科研通微信公众号,转发送积分 2482629
求助须知:如何正确求助?哪些是违规求助? 2144940
关于积分的说明 5471821
捐赠科研通 1867316
什么是DOI,文献DOI怎么找? 928181
版权声明 563073
科研通“疑难数据库(出版商)”最低求助积分说明 496574