2023 06 13 23291311 Full
2023 06 13 23291311 Full
1 Exploring the Role of Chat GPT in patient care (diagnosis and Treatment) and medical research: A Systematic Review
2
3
4
5
6
7 Ravindra Kumar Garg
8 Vijeth L Urs
9 *Akshya Anand Agrawal
10 Sarvesh Kumar Chaudhary
11 **Vimal Paliwal
12 ***Sujita Kumar Kar
13
14
15 Department of Neurology
16 King George’s Medical University
17 Lucknow, India: 226003
18
19 *Department of Surgery
20 King George’s Medical University
21 Lucknow, India: 226003
22
23 **Department of Neurology
24 Sanjay Gandhi Institute of Medical Sciences
25 Lucknow, India: 226001
26
27 ***Department of Psychiatry
28 King George’s Medical University
29 Lucknow, India: 226003
30
31
32
2
33 Correspondence
34 Ravindra Kumar Garg
35 Department of Neurology
36 King George’s Medical University
37 Lucknow, India: 226003
38 Mobile: +91 9335901790
39 Email: [email protected]
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
3
60 Abstract
61 Background
62 ChatGPT(Chat Generative Pre-trained Transformer) is an artificial intelligence (AI) based on a natural language processing tool
63 developed by OpenAI (California, USA). This systematic review examines the potential of Chat GPT in diagnosing and treating
65
66 Methods
67 In order to locate articles on ChatGPT's use in clinical practise and medical research, this systematic review used PRISMA
68 standards and conducted database searches across several sources. Selected records were analysed using ChatGPT, which also
69 produced a summary for each article. The resultant word document was transformed to a PDF and handled using ChatPDF. The
70 review looked at topics pertaining to scholarly publishing, clinical practise, and medical research.
71
72 Results
73 We reviewed 118 publications. There are difficulties and moral conundrums associated with using ChatGPT in therapeutic settings
74 and medical research. Patient inquiries, note writing, decision-making, trial enrolment, data management, decision support,
75 research support, and patient education are all things that ChatGPT can help with. However, the solutions it provides are frequently
4
76 inadequate and inconsistent, presenting issues with its originality, privacy, accuracy, bias, and legality. When utilising ChatGPT for
77 academic writings, there are issues with prejudice and plagiarism, and because it lacks human-like characteristics, its authority as
79
80 Conclusions
81 ChatGPT has limitations when used in research and healthcare. Even while it aids in patient treatment, concerns regarding
82 accuracy, authorship, and bias arise. Currently, ChatGPT can serve as a "clinical assistant" and be a huge assistance with
84
85
87
88
89
90
91
5
92 Introduction
93 ChatGPT(Chat Generative Pre-trained Transformer) is an artificial intelligence (AI) based on a natural language processing tool
94 developed by OpenAI (California, USA). ChatGPT is chat boat based technology. A chatbot is in fact a type of software creates text
95 akin to human-like conversation. ChatGPT has the capacity to respond to follow-up questions, recognise errors, debunk unfounded
96 theories, and turn down inappropriate requests. Large Language Models (LLMs), which are frequently abbreviated as LLMs, are
97 extremely complex deep-learning programmes that are capable of comprehending and producing text in a manner that is strikingly
98 comparable to that of humans. LLMs can recognise, summarise, translate, predict, and create text as well as other sorts of
99 information by using the large knowledge base they have amassed from massive datasets.1,2
100
101 The possible uses of ChatGPT in medicine is currently under intense investigation. ChatGPT is considered to have enormous
102 capability in helping experts with clinical and laboratory diagnosis to planning and execution of medical research. Another
103 significant use of ChatGPT in medical researchers is the creation of virtual assistants to physicians helping them in writing
104 manuscripts in more efficient way. Usage of ChatGPT in medical writing is considered to have associated with several ethical and
105 legal issues. Possible copyright violations, medical-legal issues, and the demand for openness in AI-generated content are a few of
106 these.3-5
107
6
108 In this systematic review we aimed to review published article and explore the potential of ChatGPT in facilitating patient care,
109 medical research and medical writing. We will also focus on ethical issues associated with usage of ChatGPT.
110 Methods
111 We performed a systematic review of published articles on ChatGPT. The protocol of the systematic review was registered with
112 PROSPERO (PROSPERO 2023 CRD42023415845).6 Our systematic review was conducted following the Preferred Reporting
114
116 We searched four databases, PubMed, Scopus, Embase, and Google Scholar. Our search was aimed at identifying all kinds of
117 articles on ChatGPT and its application in medical research, scholarly and clinical practice, published till 24 May 2023. Articles
118 related to medical education was not considered. The search item that we used was “ChatGPT”. We reviewed all kinds of
119 publications including original articles, reviews, editorial/ commentaries and even letter to the editor describing ChatGPT.
120
122 The selection of the papers that were published was done in two steps. Two reviewers (RK and VKP) reviewed the titles and
123 abstracts in the initial phase. Two reviewers (VU and SKC) then examined the entire texts of the chosen papers to determine their
7
124 eligibility. A third author (SK) settled any differences that arose between the two authors. Two reviewers (RK and VKP) assessed
125 the information available in the included publication for the suitability of the article to be included in the review. Any disagreement
126 between them was resolved by mutual agreement. If a dispute persisted, it was resolved via consultation with a third reviewer (SK).
127
128 EndNote 20 web tool (Clarivate Analytics) was used to handle duplicate records. This process was carried out by two reviewers
129 independently (RK and VKP). Any issue that arose was resolved with a discussion with another reviewer. The number of retrieved
130 and assessed records at each stage was provided in the form of a PRISMA flow chart. EndNote 20 (Clarivate Analytics) was used
132
135
137 ChatGPT was extensively used for analysing the selected records and writing this manuscript. A table was made with six columns
138 (First author/sole author, country of origin, status of peer review (peer-reviewed or preprint), title of the paper and short point wise
139 summary of full text. Short point wise summary of full text of each and every article was created with the help of ChatGPT. The
8
140 voluminous word file was then converted to a pdf file and was processed with the sister software “ ChatPDF” (OpenAI, California,
141 USA available at https://www.chatpdf.com/). Following questions were asked from ChatPDF.
142
143 1. What are potential role of ChatGPT in medical writing and research?
147 5. Can ChatGPT write text in good English and free of plagiarism?
148 6. Role of ChatGPT so far in neurological disorders related clinical practice and research.
149 7. Effectiveness and efficiency of ChatGPT in medical research and clinical settings
150 8. Potential benefits and limitations of ChatGPT in medical research and clinical applications
151 9. The ethical implications of using ChatGPT in medical research and clinical practice
152 10. Identify the gaps in the current research on ChatGPT and suggest areas for further investigation.
153 11. Provide insights into the potential future applications of ChatGPT in medical research and clinical practice
154 12. Recommendations for researchers, clinicians, and policymakers on the use of ChatGPT in medical research and clinical
155 practice
9
156
158
159 Results
160 Our data collection followed PRISMA guidelines. (Additional file 1: PRISMA checklist) The PRISMA flowchart for our systematic
161 review is shown in Figure 1. We reviewed 118 publications. ChatGPT related publications are available from across the globe.
162 There were 33 original articles and rest were commentary/ editorial, review articles, research letters or letter to the editors. Out of
163 118 articles, 18 articles were available as preprint only. Summaries of 118 articles and answers to 12 questions have been
165
166 Discussion
167 We looked into two main uses of ChatGPT: in healthcare settings and for medical writing and research. We studied 118 articles -
168 most were opinion pieces, commentaries, and reviews. Another group, Ruksakulpiwat et al, also did a similar study. They analyzed
169 six articles out of 114 that met their criteria. These articles covered a variety of ways to use ChatGPT, such as finding new drugs,
170 writing literature reviews, improving medical reports, providing medical info, bettering research methods, analyzing data, and
172 Levin et al, on the other hand, conducted an analysis of the first batch of publications about ChatGPT. They found 42 articles
173 published in 26 journals in the 69 days after ChatGPT was launched. Only one was a research article. The rest were mostly
174 editorials and news pieces. Five publications focused on studies on ChatGPT. There were no articles on its use in Obstetrics and
175 Gynecology.In terms of where these articles were published, Nature was the top journal. Radiology and Lancet Digital Health came
176 next. The articles mostly discussed the quality of ChatGPT's scientific writing, its features, and its performance. Some also talked
177 about who should get credit for the work and ethical concerns. Interestingly, when comparing the articles that described a study to
178 the others, the average impact factor (a measure of the influence of a journal) was significantly lower for the study articles.8
179
180 In our review, we identified several potential advantages of using ChatGPT in the medical field. It appears to enhance productivity
181 and expedite research workflows by aiding in data organization, assisting in the selection of trial candidates, and supporting overall
182 research activities. Furthermore, ChatGPT's capacity to review manuscripts and contribute to editing may potentiate the efficiency
183 of academic publishing. Beyond the scope of research, it could also prove beneficial for patient education, fostering scientific
184 exploration, and shaping clinical decision-making. However, we also need to consider certain limitations and ethical concerns
185 associated with the use of ChatGPT. The model, as sophisticated as it is, lacks the capability to offer comprehensive diagnoses
186 and cannot replace the human qualities inherent to medical practice. Ethical issues also arise, specifically in relation to potential
187 biases in the machine learning model and potential breaches of privacy. Moreover, while ChatGPT can process and generate
11
188 information, it might not exhibit the level of originality, creativity, and critical thinking that are often required in the medical field.
189 However, the use of ChatGPT in producing scholarly articles is raising questions in the academic publishing. While these tools can
190 greatly enhance the clarity and fluency of written material, it is crucial that human oversight is maintained throughout the process.
191 This is because AI can potentially produce content that is authoritative-sounding, yet it might be inaccurate, incomplete, or biased.
192 Incorrect GPT-4 responses, known as "hallucinations," can be harmful, particularly in the field of medicine. Therefore, it is essential
193 to check or validate GPT-4's output. ChatGPT can generate references to made-up research publications. Therefore, authors must
194 thoroughly check and modify the output of these tools. Furthermore, it is not appropriate to recognize AI or AI-assisted tools as
195 authors or co-authors in the by-line of publications. Instead, their use should be transparently acknowledged within the manuscript.
196 For example, according to Elsevier's policy on AI for authors, the responsibility and accountability for the work ultimately still lie with
197 the human authors, despite any technological assistance they may have received.9-11
198
199 In conclusion, ChatGPT has a great potential. Its full potentials are still evolving. ChatGPT as a source of information can not be
200 trusted, many ethical issues are associated with it. Certainly, ChatGPT can be credited with authorship. However, ChatGPT is
201 certainly a good clinical assistant. ChatGPT is nowhere near to replace human brain.
202
203
12
204 Acknowledgment
205 The concept, data collection analysis, writing, and reporting of this article were solely done by authors. ChatGPT was extensively
207
210
213
215 None
216
217
218
219
13
220 References
221 1. OpenAI. (2023). Conversational AI Model for Medical Inquiries. ChatGPT. https://www.openai.com/chatgpt.
222
223 2. Lee P, Bubeck S, Petro J. Benefits, Limits, and Risks of GPT-4 as an AI Chatbot for Medicine. N Engl J Med. 2023 Mar
225
226 3. Bommasani R, Liang P, Lee T. Holistic Evaluation of Language Models. Ann N Y Acad Sci. 2023 May 25. doi:
228
229 4. Waisberg E, Ong J, Masalkhi M, Kamran SA, Zaman N, Sarker P, Lee AG, Tavakkoli A. GPT-4: a new era of artificial
230 intelligence in medicine. Ir J Med Sci. 2023 Apr 19. doi: 10.1007/s11845-023-03377-8. Epub ahead of print.
231
232 5. Ong CWM, Blackbourn HD, Migliori GB. GPT-4, artificial intelligence and implications for publishing. Int J Tuberc Lung Dis. 2023
234
14
235 6. Garg RK, Paliwal V, Kar SK, Urs VL, Chaudhary SK. Exploring the Role of Chat GPT in patient care (diagnosis and Treatment)
236 and medical research: A Systematic Review. PROSPERO 2023 CRD42023415845 Available
238
239 7. Ruksakulpiwat S, Kumar A, Ajibade A. Using ChatGPT in Medical Research: Current Status and Future Directions. J Multidiscip
241
242 8. Levin G, Brezinov Y, Meyer R. Exploring the use of ChatGPT in OBGYN: a bibliometric analysis of the first ChatGPT-related
243 publications. Arch Gynecol Obstet. 2023 May 24. doi: 10.1007/s00404-023-07081-x. Epub ahead of print.
244
245 9. Anonymous. Tools such as ChatGPT threaten transparent science; here are our ground rules for their use. Nature. 2023
247
248 10. Hill-Yardin EL, Hutchinson MR, Laycock R, Spencer SJ. A Chat(GPT) about the future of scientific publishing. Brain Behav
250
15
251 11. Thorp HH. ChatGPT is fun, but not an author. Science. 2023 Jan 27;379(6630):313. doi: 10.1126/science.adg7879. Epub 2023
253
254
255
256
257
258
259
260
261
262
263
264
265
266
16
267 Legend
268 Figure 1: The study's PRISMA flow diagram shows how articles are selected for this systematic review.
269
270
271
272
273
274
275
276
277
278
279
280
281
282
17
283 Table-1: Summaries of full text of 118 selected article. Summaries were created with the help of ChatGPT. PROMPT, that
284 given was “Provide point wise summary” .
285
Reference Country Type of Status of Title of the paper Main findings of the paper
publication peer
review
Ali and Djalilian India Editorial Peer Readership awareness The question of whether a chatbot like ChatGPT can be regarded as an author
2023 reviewed series–paper 4: gives rise to ethical dilemmas concerning accountability for the content produced
Chatbots and ChatGPT- and the accuracy of source attributions.
ethical considerations in
scientific publications
Ali et al 2023 United Commentary Peer Using ChatGPT to write CHATBOTS, such as ChatGPT, were employed by researchers to generate a
Kingdom reviewed patient clinic letters variety of clinical communication scenarios that were designed to be
comprehensible for individuals at or below a sixth-grade level (approximately 11-12
years of age).
Alser and Egypt Commentary Peer Concerns with the ChatGPT, a language model, was credited authorship in four medical papers. It
Waisberg 2023 reviewed Usage of ChatGPT in met one of four ICMJE authorship criteria by contributing to writing parts of the
Academia and papers. Plagiarism was found in its contributions. Unclear sources of information
Medicine: A and bias due to tuning also limit its use in scientific contexts.
Viewpoint
Anderson et al Germany and Editorial Peer AI did not write this The ability of AI to produce academic papers has been tested and found wanting.
2003 other reviewed manuscript, or did The papers produced by AI lacked fresh ideas, in-depth knowledge of the subject,
European it? Can we trick the AI and precise referencing. There were problems with equality, accuracy, poor AI
countries text detector into detection, unethical content creation, and plagiarism. To maintain the quality of
generated texts? The scientific literature, the application of AI tools needs to be carefully studied. AI
potential future of misuse can be avoided with the aid of current plagiarism-detection tools and
ChatGPT and AI in possible cutting-edge software for AI-generated text detection.
Sports & Exercise
Medicine manuscript
generation
Arun Babu and India Letter to the Peer Using artificial The use of AI-generated content in medical journals raises ethical concerns
Sharmila 2023 editor reviewed intelligence chatbots regarding accuracy, bias, authorship, and disclosure, necessitating new guidelines
like 'ChatGPT' to draft and standards for publication integrity.
articles for medical
journals - Advantages,
limitations, ethical
concerns and way
forward
Asch USA Commentary Peer An Interview with ChatGPT has potential in healthcare, assisting with tasks like virtual assistance,
2023 Interview reviewed ChatGPT About Health documentation, research, education, and patient engagement. Concerns such as
and chat Care privacy, bias, regulation, and ethics must be addressed. It can both improve and
with reduce equity. Advancements should focus on data quality, NLP, integration,
18
2023 reviewed Practice—a New Kid on making, preoperative education, medical record transcription, surgical planning,
the Block and medical education. However, limitations include potential misinformation, bias,
reliance on outdated data, copyright concerns, technical issues, and lack of
personalization. Proper evaluation, oversight, and addressing legal and ethical
considerations are necessary for optimal use.
Biswas 2023 USA Commentary Peer ChatGPT and the future Article introduces AI and ChatGPT.
reviewed of medical writing
Boßelmann et al USA Commentary Peer Are AI language models Artificial Intelligence (AI) can significantly improve epilepsy diagnosis and
2023 reviewed such as ChatGPT ready treatment. Examples of epilepsy-related AI research include EEG data analysis,
to improve the care of MRI-based lesion detection, and Clinical Decision Support Systems. While AI
individuals with language models like ChatGPT show promise in improving patient care, they
epilepsy? require rigorous testing, validation, and ethical consideration, along with the
expertise of medical professionals.
Brainard 2023 USA Commentary Peer Journals take up arms Publishing directors and journal managers are concerned about the impact of AI-
reviewed against AI-written text powered chatbots like ChatGPT on scholarly literature. These tools can produce
accurate-sounding, AI-generated reports and scientific manuscripts, raising issues
of accuracy, authorship, and potential flooding of AI-produced manuscripts.
Publishers are formulating policies requiring disclosure of AI tool use and are
exploring technology to detect synthetic text.
Cahan and Switzerland Editorial Peer A conversation with Computational biology and systems biology have made significant contributions to
Treutlein 2023 reviewed ChatGPT on the role of stem cell research by enabling the analysis of large datasets, identifying patterns
computational systems and trends, optimizing stem cell culture conditions, and creating detailed models of
biology in stem cell stem cell behavior. Challenges include the need for extensive data collection and
research processing, but advancements in data standardization and accessibility have
facilitated data reuse and meta-analyses. Integrating computational techniques with
stem cell research holds great promise for advancing our understanding and
applications of stem cells.
Nasrallah 2023 Editor-in- Editorial Peer A ‘guest editorial’ … The article discusses the introduction of ChatGPT, an AI program with potential
Chief Guest reviewed generated by societal impact. It highlights concerns about accuracy and references, tests
Current editorial ChatGPT? ChatGPT's abilities, and reflects on the era of AI-generated articles.
Psychiatry generated
by ChatGPT
Cascella et al Italy Review Peer Evaluating the feasibility This study examines ChatGPT's possible uses in clinical and research settings,
2023 reviewed of ChatGPT in including supporting clinical practise, doing scientific work, preventing potential
healthcare: an analysis abuse, and debating public health issues. The study highlights ChatGPT's
of multiple clinical and shortcomings and emphasises the need for careful use and further development to
research scenarios avoid dangers, such as disinformation and misuse, even if it shows promise in a
number of areas.
Chen et al 2023 USA Original Preprint The utility of ChatGPT This study found that while ChatGPT correctly aligned with National
for cancer treatment Comprehensive Cancer Network (NCCN) guidelines in 98% of cases, it also
information provided partially non-concordant advice in 34.3% of them. Therefore, it concluded
that ChatGPT is not entirely reliable for cancer treatment advice, urging patients
and clinicians to recognize its limitations.
20
Cheng et al 2023 China Letter to the Peer- ChatGPT/GPT-4: The potential application of ChatGPT/GPT-4 in surgical oncology, from aiding in
editor reviewed Enabling a new era of clinical trial design, case management, data analysis, to preoperative preparations
surgical oncology and intraoperative processes, is discussed in this study. Despite potential
challenges and ethical concerns, the authors advocate for responsibly harnessing
this AI technology to improve surgical oncology outcomes.
Chervenak et al USA Original Peer- The promise and peril ChatGPT, tested against established fertility sources, provided responses
2023 reviewed of using a large comparable in quality to the CDC's FAQs and scored highly on validated fertility
language model to knowledge surveys. Although it effectively reproduced missing facts from the
obtain clinical American Society for Reproductive Medicine's committee opinion, its clinical utility
information: might be limited by occasional incorrect information and lack of source citation.
ChatGPT performs
strongly as a fertility
counseling tool with
limitations
Cifarelli and USA Commentary Peer- Large language model However, ethical considerations arise when utilizing ChatGPT in scholarly
Sheehan 2023 reviewed artificial intelligence: the publishing. These encompass potential bias in training data and the perpetuation of
current state and future prejudice, concerns about authorship, copyright, and plagiarism, implications for
of ChatGPT in neuro- citation practices and the "Matthew Effect," and impacts on academic job
oncology publishing expectations, tenure, and promotion.
D'Amico et al USA Editorial Peer- I asked a ChatGPT to This editorial discusses the potential and challenges of incorporating AI chatbots in
2023 reviewed write an editorial about neurosurgery for data collection, patient care, and communication. However, issues
how we can incorporate like potential misinformation, privacy, ethical concerns, bias, legal liabilities, content
chatbots into validity, and effectiveness need addressing. It emphasizes human responsibility in
neurosurgical research verifying machine-generated content for moral and ethical standards. Neurosurgery
and patient care… should lead in responsibly integrating these AI technologies.
Darkhabani et al Turkey Review Peer- ChatGPT and This paper explores the role of AI language model ChatGPT in clinical medicine,
2023 reviewed autoimmunity - A new specifically autoimmunity, discussing its capabilities, limitations, and potential cyber
weapon in the risks. It highlights the importance of continuous evaluation, as ChatGPT and similar
battlefield of knowledge technologies evolve rapidly, emphasizing the need for healthcare professionals to
stay abreast of these developments.
Dave 2023 UK Commentary Peer- Plagiarism software Turnitin, the academic integrity firm used by 98% of UK universities, can now
reviewed now able to detect detect plagiarism in content generated by artificial intelligence systems like
students using ChatGPT. This development aims to maintain academic integrity in the face of
ChatGPT advancements in AI technology, preventing dishonest use of AI-generated work.
Dave et al 2023 UK Review Peer- ChatGPT in medicine: ChatGPT can be a valuable tool for medical research and patient care, providing
reviewed an overview of its support in data analysis and diagnosis generation. However, it should supplement,
applications, not replace, human expertise. Ethical considerations, like data privacy and
advantages, limitations, accuracy of recommendations, need to be addressed. Future prospects include
future prospects, and integration with other AI technologies, prioritizing a balanced, cautious, and
ethical considerations ethically mindful approach.
Day et al 2023 Canada Original Peer- A Preliminary ChatGPT provide citations as well although recent scrutiny revealed that these
article reviewed Investigation of Fake references are often incorrect or non-existent, according to a study. Despite the
Peer-Reviewed impressive capabilities of the chatbot, the presence of fake references and citations
Citations and is a concern for both academic research and student academic integrity. While
References Generated these inaccuracies present a problem for those who might use the bot for sourcing,
by ChatGPT they could serve as a useful tool for identifying academic misconduct or as prompts
for further research. Despite these issues, the chatbot has been utilized effectively
for generating course materials in certain subject areas, suggesting potential for its
continued use in education. However, the tool requires subject expertise and its
use should be approached with caution.
de Oliveira and Brazil Editorial Peer- The future of Pediatric The use of AI like ChatGPT in medical writing is revolutionary, but has ethical
Ballestero 2023 reviewed Neurosurgery and implications and potential for errors. AI's usage should be transparent, and authors
ChatGPT: an editor's should ensure their publications' scientific integrity.
perspective
De Vito 2023 Argentina Editorial Peer- Artificial intelligence and This editorial discusses the implications of using AI, particularly OpenAI's GPT-3, in
reviewed chatGPT. Would you scientific and academic writing. While acknowledging the benefits of AI tools in
read an artificial author? producing coherent text, it raises concerns about authenticity, potential misuse, and
the need for transparency. It also stresses the importance of human responsibility
in ensuring the accuracy of scientific content and the ethical use of AI technology in
academia.
Dergaa et al 2023 Multiple Original Peer- From human writing to This study explores the benefits and potential threats of NLP technologies like
22
countries article reviewed artificial intelligence ChatGPT in academic writing and research. While these tools can enhance
generated text: efficiency, concerns arise about the authenticity and credibility of work. The study
examining the underscores the need for ethical considerations, transparency, and human
prospects and potential intelligence in their use to maintain academic integrity.
threats of ChatGPT
in academic writing
Donato et al 2023 Portugal Editorial Peer The Transparency of This article discusses ethical implications of large language models (LLMs) like
reviewed Science with ChatGPT ChatGPT in scientific publishing. It highlights concerns about authorship,
and the Emerging transparency, and integrity in utilizing AI tools. The need to declare AI's role and
Artificial Intelligence limit its misuse is emphasized. It also discusses future prospects of detecting AI-
Language Models: generated text and potential applications of LLMs in non-English language
Where Should Medical contexts.
Journals Stand?
Dunn et al 2023 USA Original Peer Artificial intelligence- This letter discusses a study comparing the quality and readability of dermatology-
article reviewed derived dermatology based case reports produced by human authors and AI models like ChatGPT.
case reports are Findings suggest that AI-generated reports were often indistinguishable from
indistinguishable from human-created ones, though some limitations were noted. AI detection tools varied
those written by in effectiveness, raising ethical implications about the reliability, accuracy, and
humans: A single- transparency of AI-generated content. The authors argue for the need to maintain
blinded observer study integrity of AI-produced content, and for editorial teams to consider policies around
the use of AI and Large Language Models.
Fatani 2023 Saudi Arabia Review Peer ChatGPT for Future This narrative review discusses the application and implications of AI language
reviewed Medical and Dental model, ChatGPT, in medical and dental research. ChatGPT can assist in writing
Research scientific papers, summarizing data, and translating languages. While AI can
enhance clinical workflow and help generate quick responses, it's cautioned that
over-reliance could lead to papers lacking critical thinking. Ethical concerns such as
plagiarism and data integrity are also highlighted. The review concludes that while
ChatGPT can support research and potentially transform clinical medicine, there
are limitations and potential risks that need to be carefully managed.
Galland 2023 France Editorial Peer Chatbots and internal In November 2022, OpenAI® made headlines by releasing version 3.5 of its AI
reviewed medicine: Future ChatGPT, a revolutionary conversational AI capable of generating text on any
opportunities and topic, including in the medical field. It was joined in 2023 by other advanced
challenges generative AIs such as Google's BardAI and image generators like Dall-E 2 and
Mid Journey. The use cases in healthcare are increasing, and chatbots could be
valuable tools for prevention, appointment scheduling, symptom information,
administrative tasks, and more. However, concerns about data privacy, user safety,
and accuracy of information arise with the use of chatbots in healthcare. The
adoption of these technologies by patients may outpace the establishment of
scientific consensus on their use. The article emphasizes the need for caution,
evaluation, and defining the indications and limitations of their utilization in the
medical field.
Gandhi India Letter to the Peer ChatGPT: roles and The article discusses the advancements in AI, particularly ChatGPT, an advanced
23
Periaysamy et al editor reviewed boundaries of the new language model developed by OpenAI. It highlights its potential in medical
2023 artificial intelligence tool education and health research. ChatGPT can provide answers and information for
in medical education medical students, assist researchers in writing and reviewing content, and even
and health research - generate scientific abstracts. However, it also raises concerns about potential
correspondence misuse, integrity of generated content, and ethical considerations regarding
authorship in research papers involving AI. Proper guidelines and reporting
standards are recommended to ensure responsible and transparent use of AI tools
in the medical field.
Gao et al 2023 USA Original Peer- Comparing scientific The article assesses ChatGPT's capacity to produce scientific abstracts, a sizable
article reviewed abstracts generated by language model. Although neither was flawless, both human reviewers and AI
ChatGPT to original output detectors were able to recognise some created abstractions. The created
abstracts using an abstracts were in the same range as actual abstracts but contained made-up
artificial intelligence numbers. The study is aware of its limitations, including the sample size and the
output detector, type of cues employed. The technology has the potential to be used in both ethical
plagiarism detector, and and immoral ways, such as to fabricate research or aid in research writing. The
blinded human biases in training data are underlined, as is the requirement for more investigation
reviewers into potential biases.
Goedde et al 2023 Germany Original Preprint ChatGPT in medical This concise review focuses on ChatGPT (Chat Generative Pre-trained
article literature-a concise Transformer) and its role in medical literature. The study analyzed the literature
review and SWOT published on ChatGPT from December 2022 to March 2023 and conducted a
analysis SWOT analysis. The findings highlight the strengths, weaknesses, opportunities,
and threats associated with ChatGPT. The review emphasizes the need for further
research, regulations, and policies to address the potential of ChatGPT in medical
literature.
Gordijn and Have Ireland Editorial Peer- ChatGPT: evolution or This editorial explores the implications of large language models (LLMs),
2023 reviewed revolution? specifically ChatGPT by OpenAI, on academic publishing and research. It
questions whether the development of LLMs will necessitate new editorial policies
to address the challenges they pose. Viewing ChatGPT as either evolutionary or
revolutionary, the article suggests the utility of the AI model as a research tool may
be limited due to deficiencies. It acknowledges the potential for LLMs to improve
research but also highlights potential risks such as plagiarism and factual
inaccuracies. Current editorial policies may already handle these issues, but further
review and adjustments could be necessary.
Gottlieb et al 2023 USA Commentary Peer- ChatGPT and AI and machine learning are significantly impacting healthcare and research,
reviewed conversational artificial enhancing precision medicine through big data analysis and personalized
intelligence: Friend, foe, treatment plans.
or future of research? Large Language Models (LLMs) like ChatGPT serve as conversational interfaces,
aiding in patient communication and assisting researchers in content generation.
Advantages include accelerating content creation, language translation, simplifying
complex topics, synthesizing literature, creating abstracts, identifying limitations,
and disseminating knowledge.
Potential harms encompass limitations in the extant knowledge of LLMs, potential
misinterpretation of study methods/results, overconfidence in AI-generated content,
24
ChatGPT agrees with this sentiment, stating it doesn't possess the ability to
consent to co-authorship, but can assist with writing and editing.
When applying for full registration, students are obligated to disclose any fitness to
practice investigations to the GMC.
Mein highlights that some students have been taken aback when contacted by the
GMC about undisclosed investigations.
To help students avoid plagiarism in the "age of AI," the MDU has provided tips that
address common errors such as "para-plagiarizing" and insufficient citation.
Gravel et al 2023 Canada Original Preprint Learning to fake it: The objective of this study was to assess the quality and suitability of responses
article limited responses and and references provided by ChatGPT when addressing medical questions. The
fabricated references study selected 20 diverse medical questions derived from recent research articles
provided by ChatGPT published in high-impact factor medical journals. These questions were related to
for medical questions the main objectives of the articles or framed within a broader context to ensure a
comprehensive range of references.
The study followed a specific methodology where the questions were posed to
ChatGPT without imposing any word limit or constraint. After receiving the
response, a follow-up query was made to request references, with only the first
three references being considered for analysis. To ensure domain expertise, the
corresponding authors of the selected articles were invited to rate the responses.
The primary outcomes measured in the study were the appropriateness of the
references and the quality of the responses.
The findings of the study revealed that the majority of references provided by
ChatGPT were fabricated. These references appeared credible, featuring authors
with previous publications or affiliations with reputable organizations. However,
69% of the references were found to be nonexistent. Among the 59 references
included in the analysis, only 18 were authentic, although they contained minor or
major citation errors. The remaining references were completely fictional.
Additionally, the study reported that 95% of the references listed authors with prior
publications or had affiliations with recognized organizations.
To summarize, the study revealed that the references provided by ChatGPT were
predominantly fabricated, and the quality of the responses varied, including
instances of major factual errors. These findings raise concerns regarding the
reliability and accuracy of the references and responses generated by ChatGPT
when addressing medical questions.
Guo et al 2023 Canada Original Preprint neuroGPT-X: Towards This study enhanced OpenAI's GPT-3 model for answering questions about
article an Accountable Expert vestibular schwannoma. The context-enriched GPT model provided faster and non-
Opinion Tool for inferior responses compared to expert neurosurgeons. The model's responses
Vestibular were often rated higher and included in-text citations and references. However,
Schwannoma expert surgeons expressed concerns about the model's reliability in addressing
nuanced and controversial aspects of management. The study also introduced
neuroGPT-X, a chat-based platform for clinical support with accurate and reliable
information.
Gurha et al 2003 USA Commentary Peer- ChatGPT and other An AI chatbot or AI writer should not be listed as an author in scientific publications
reviewed artificial intelligence because authors provide nuanced interpretation and insight beyond synthesizing
chatbots and sentences.
biomedical writing Authors cannot claim ownership of text generated by chatbots and should avoid
plagiarism by acknowledging their use.
Chatbots are limited in generating in-depth and critical scientific writings and lack
the essential insight required for review articles, commentaries, or perspectives.
Large language models (LLMs) have potential in scientific writing but should be
carefully evaluated for validity and edited for incorrect statements.
Chatbots have various biomedical and clinical applications but require expert
review and verification.
LLM-based programs face challenges in extracting reliable data from biased and
low-quality sources.
As programs advance, commercialization may occur, but scientists should maintain
access to platforms for synthesizing and extracting advanced information.
JCA editors prioritize articles with nuanced insight and new understanding.
Hurley 2023 USA Commentary Peer- Your AI Program Will ChatGPT, the language model developed by OpenAI, is creating ripples in the
reviewed Write Your Paper Now: academic world, with several journals considering guidelines for its usage. Experts
Neurology Editors on note the AI's remarkable writing ability and its potential to generate data and
Managing Artificial references, which raises concerns about authenticity. While some editors suggest
Intelligence there could be a place for ChatGPT to assist with language barriers and manuscript
Submissions restructuring, there's consensus that the AI should not be credited as an author,
and its usage should be disclosed. Additionally, experts emphasize the need for
human oversight and responsibility for content generated with AI's help.
Haemmerli et al Switzerland Original Preprint ChatGPT in glioma In this study, the performance of ChatGPT, an AI-based chatbot, was assessed in
2023 article patient adjuvant therapy providing treatment recommendations for glioma patients within a multidisciplinary
decision making: ready decision-making context. ChatGPT demonstrated accuracy in identifying cases as
to assume the role of a gliomas; however, it faced challenges when specifying tumor subtypes. The
doctor in the tumour treatment recommendations and regimens generated by the tool were generally
27
board? rated as good, while its ability to consider functional status received a moderate
rating. Experts found ChatGPT's involvement in the CNS Tumor Board to be
valuable, recognizing its potential for improvement and learning.
Harskamp and Netherlands Original Preprint Performance of The AMSTELHEART-2 study evaluated the performance of the ChatGPT language
Clercq 2023 article ChatGPT as an AI- model in answering cardiovascular trivia questions and interpreting case vignettes.
assisted decision The model achieved an overall accuracy of 74% in answering multiple-choice
support tool in questions and provided appropriate advice for patients reaching out to primary
medicine: a proof-of- care. However, it had limitations in addressing medical questions from physicians
concept study for seeking expert consultation. The study's strengths include using various
interpreting symptoms approaches to evaluate ChatGPT, while its limitations include a small sample size.
and management of Further refinements of the model are needed before considering its use in real-life
common cardiac scenarios.
conditions
(AMSTELHEART-2)
Hill-Yardin et al Australia Viewpoint- Peer- A Chat(GPT) about the The use of AI-generated software, such as ChatGPT, in academic publishing has
2023 Commentary reviewed future of scientific sparked discussions and debates.
publishing The generated text can be accurate, logical, and grammatically correct, but it lacks
a distinct "voice" and is generic.
The writing style of the generated text may resemble the bland and formulaic style
often found in scientific articles.
Introducing more diversity in writing styles and discussing mistakes and
unsupported hypotheses could help differentiate human writing from AI-generated
writing.
AI may eventually learn individual writing styles and inject more nuanced diction.
The use of AI and predictive text software raises questions about language
patterns, vocabulary, and linguistic diversity.
ChatGPT's generated text is too shallow for the detailed knowledge and
interpretation required in neuroscience.
The lack of accountability and transparency in AI-generated text is a fundamental
concern for its use in academic publishing.
Hirani et al 2023 USA Commentary Peer- Experimenting with The functionality of ChatGPT in academic medical publishing was examined,
reviewed ChatGPT: Concerns for revealing the potential for fabrication of inaccurate research.
Academic Medicine The lack of transparency and accountability in ChatGPT's output raised concerns
and prompted guidelines from publishing companies and journal editors.
Bias is a significant problem as well because word embeddings, which are used in
ChatGPT training, introduce biases originating from humans into the original data.
Algorithmic bias in medical AI systems, which also leads to care disparities, may
have a negative impact on clinical decision-making.An analysis uncovered biases
in ChatGPT's responses, such as professions and jobs that are linked with
particular genders. Physicians should advocate for transparency, regulation, and
accountability in LLM training datasets and algorithms to address the issues of
fabrication and bias.
28
Homolak Croatia Commentary Peer- Opportunities and risks The release of ChatGPT has generated excitement about the potential of AI in
2023 reviewed of ChatGPT in various fields, including academic publishing.
medicine, science, and AI has the potential to revolutionize healthcare but is unlikely to replace physicians
academic publishing: a due to limitations in context and nuance.
modern Promethean Ethical issues surround conversational AI in medical practice, including biased data
dilemma sets and accountability for mistakes.
ChatGPT's ability to generate scientific manuscripts raises concerns about
reliability and the potential for misinformation.
Acknowledging limitations and addressing ethical challenges are necessary before
implementing AI in publishing.
While AI has the potential to improve efficiency, a mindful approach and open
debate about risks and benefits are essential.
Hosseini et al USA Original Preprint ChatGPT versus the A hybrid panel discussion about the use of ChatGPT in education, research, and
2023 article neurosurgical written healthcare was conducted.
boards: a comparative 420 responses were received from 844 participants, with a response rate of 49.7%.
analysis of artificial 40% of the audience had used ChatGPT, with more trainees than faculty.
intelligence/machine Interest in wider application was higher among ChatGPT users.
learning performance Greatest uncertainty was around its use in education.
on neurosurgical Varied perspectives were present among different roles (trainee, faculty, staff).
board–style questions The need for further discussion about LLM usage was highlighted.
A thoughtful, measured approach in adopting ChatGPT was suggested to reduce
potential risks.
Howard et al 2023 UK Commentary Peer ChatGPT and ChatGPT has been noted for its potential in medical scenarios, stimulating urgent
reviewed antimicrobial advice: the discussions in the medical community.
end of the consulting Despite its limitations, it's considered as capable as an average human physician in
infection doctor? answering some open-ended medical queries.
The study tested ChatGPT in eight hypothetical infection scenarios.
It recognized natural language effectively but struggled with complex situational
aspects.
Its responses were coherent, clear, and reflective of its information sources.
Some inconsistencies and erroneous advice emerged on repeated questioning.
Deficits in situational awareness, inference, and consistency are major barriers to
ChatGPT's clinical implementation.
Despite no access to specific medical databases, it provides compelling responses.
The authors propose a modifiable qualitative framework for future improvements.
Hsu et al 2023 Taiwan Original Preprint Plagiarism, Quality, and This study aimed to evaluate the quality and accuracy of medical abstracts
article Correctness of generated by ChatGPT, specifically the ChatPDF version, using a selection of 20
ChatGPT-Generated vs randomised controlled trials in the field of psychiatry. Two types of abstracts,
Human-Written Abstract structured and unstructured, were considered.
for Research Paper
The generated abstracts were analyzed in terms of similarity, plagiarism, AI-
content, and subheading proportion. They were also evaluated by five experts in
psychiatry for quality, using a Likert scale. Further, the validity of the conclusions
29
The findings showed a similarity (duplicate content) between the original and the
generated abstracts of 16.35%. The plagiarism percentage was 18.75%, with
11.20% being self-plagiarism. After adjusting for self-plagiarism, the actual
plagiarism percentage was 7.55%. The AI-content percentage was notably higher
in unstructured abstracts (75.58%) compared to structured ones (31.48%).
The quality of generated abstracts was found to be lower than the original
abstracts, especially for the unstructured format. Interestingly, a structured format
and a higher H-index were associated with higher quality scores in generated and
original abstracts, respectively.
Lastly, while the ability of experts to accurately identify original authors was fairly
low (40%) for structured abstracts, it was significantly higher (73%) for unstructured
ones. However, a noteworthy concern was that 30% of the abstracts generated by
ChatGPT presented incorrect conclusions.
LLMs can assist in generating drafts for study ideas, research protocols,
manuscripts, grant proposals, instructional materials, and patient education
materials. They can also help improve text and rectify errors for non-native
speakers.
Integration of LLMs into electronic health record systems could enable information
extraction from patient records and other data repositories, automating data
31
Advanced LLMs could find application in clinical workflows for evaluating patient
information and generating patient management plans, leading to more efficient
patient care and improved outcomes.
However, there are limitations and risks associated with using LLMs. These include
the possibility of "neural hallucinations," where the model generates incorrect or
nonsensical text, and the potential introduction of bias into the model's output.
Establishing precise criteria and guidelines for LLM usage is crucial to ensure their
trustworthiness and security in surgical science, particularly when considering the
ethical and legal implications of their use in patient management.
Despite these limitations, LLMs have the potential to be valuable tools in surgical
science and clinical practice by augmenting human expertise rather than replacing
it.
There were minimal differences observed in word count and readability grade level
between the responses from NCI and ChatGPT. The study suggests that ChatGPT
provides accurate information on common cancer myths and misconceptions
without disseminating misinformation or harmful information.
Further research is needed to assess the accuracy of other chatbots and AI-driven
systems, their performance in addressing diverse claims about cancer, and the
development of an ideal infrastructure to monitor and ensure the accuracy of
cancer information online.
Juhi et al 2023 India Original Peer The Capability of The study aimed to investigate the effectiveness of ChatGPT in predicting and
article reviewed ChatGPT in Predicting explaining common drug-drug interactions (DDIs).
and Explaining A total of 40 DDI pairs were used to converse with ChatGPT using a two-stage
Common Drug-Drug question approach.
Interactions The first question asked whether two specific drugs can be taken together, and the
second question asked why those drugs should not be taken together.
The responses from ChatGPT were checked by two pharmacologists and
categorized as "correct" or "incorrect." The "correct" responses were further
classified as "conclusive" or "inconclusive."
Among the 40 DDI pairs, one answer was incorrect in the first question, and one
answer was wrong in the second question.
For the first question, 19 answers were classified as conclusive and 20 as
inconclusive. For the second question, 17 answers were conclusive, and 22 were
inconclusive.
Using Flesch reading ease and Flesch-Kincaid grade level, the readability ratings
of the answers were assessed. The answers were found to have substantially
higher grade levels than anticipated for a fictitious sixth grade, making them rather
challenging to read.
According to the study, ChatGPT can be used to predict and explain DDIs to a
limited extent. Patients without rapid access to medical facilities may receive some
guidance from it, but it may also provide insufficient information. It need further
work before patients might possibly use it to comprehend DDIs.
Kaneda Japan Commentary Peer- In the Era of Prominent A large language model (LLM) called ChatGPT, created by OpenAI, has become
2023 reviewed AI, What Role Will well-known for producing conversational responses.
Physicians Be Expected An AI chat application called Chat Doctor has proven to be quite accurate at
to Play? prescribing prescriptions based on talks between patients and doctors.
The ability of LLM models to aid patients and medical personnel in the sphere of
healthcare is underlined.
As AI models have limits in offering the most recent research, ethical judgement,
33
and clinical intuition, doctors will continue to play an important role in healthcare
settings. Collaboration between medical professionals, administrators, and
stakeholders is crucial to advance healthcare reform and effectively utilize AI.
Kim Republic of Original Preprint Search for Medical The study evaluated the ability of ChatGPT, an artificial intelligence chatbot, to
2022 Korea Information and provide medical information and treatment options for subacromial impingement
Treatment Options for syndrome (SIS).
Musculoskeletal ChatGPT was able to generate responses that included correct definitions,
Disorders through an prevalence, and risk factors of SIS based on input messages.
Artificial Intelligence It also provided information on symptoms, diseases with similar symptoms, and
Chatbot: Focusing on orthopedic tests related to SIS.
Shoulder Impingement ChatGPT suggested treatment options such as rest, medication, physical therapy,
Syndrome and potential surgical intervention.
However, some of the information provided by ChatGPT was biased or
inappropriate depending on the individual's condition, indicating the limitations of
relying solely on AI chatbots for accurate medical information.
The study concluded that while ChatGPT can offer overall helpful information about
SIS, caution is necessary, and it still falls short in providing accurate and
individualized medical information and treatment options.
With further advancements in natural language processing technology, it is
expected that AI models like ChatGPT can deliver more detailed and precise
medical information in the future.
Kim 2023 China Original Peer Using ChatGPT for The use of AI chatbots as authors in scientific articles is an ethical issue and has
reviewed language editing in raised concerns in scientific journals.
scientific articles ChatGPT, developed by OpenAI, is a popular AI-powered chatbot available for free
to anyone.
The study submitted questions to ChatGPT regarding the effect of streptozotocin-
induced diabetes on facial bone growth in rats.
ChatGPT provided a detailed response with references to studies, but upon
verification, all the references were found to be fake.
ChatGPT's performance in editing English grammar was found to be excellent.
Language editing services like ChatGPT can be advantageous for authors, but they
should not be listed as co-authors in scientific articles.
Koo China Letter to the Peer- The importance of ChatGPT has several helpful applications in the writing process, including
2023 editor reviewed proper use of ChatGPT paraphrasing difficult sentences, translation, spell-checking, grammar correction,
in medical writing improving clarity and flow of text, generating outlines and abstracts, and formatting
references.
The author of the article in Radiology found that ChatGPT could write an entire
article based on inputted headings and subheadings.
The potential risks of using AI in the writing process were discussed, emphasizing
the importance of proper and ethical use of ChatGPT.
ChatGPT is a versatile tool that can streamline the writing process for researchers,
but it should be used responsibly to maintain ethical standards.
Kumar 2023 Ireland Original Peer- Analysis of ChatGPT he study aimed to evaluate the effectiveness of ChatGPT, an AI chatbot, in the
34
article reviewed tool to assess the context of academic writing within the field of biomedical sciences. Five random
potential of its utility for topics were inputted into ChatGPT, and the study assessed factors such as
academic writing in response time, content quality, and reliability. The responses generated by
biomedical domain ChatGPT were exported to a Word file and analyzed for originality using Urkund
software.
The findings revealed that ChatGPT had a fast response rate, producing 300-500
words of text in under 2 minutes. However, the content fell short of the expected
quality and depth typically found in academic writing. Limitations observed included
deficiencies in word count, referencing errors, and a lack of academic merit.
In conclusion, while ChatGPT has the potential to serve as a useful tool for training
and improving academic writing skills, it cannot replace human intelligence. When
used appropriately and under academic mentoring, it can assist in refining human
capabilities. Nevertheless, it is crucial to acknowledge the limitations of ChatGPT
and complement its usage with human expertise to ensure the production of high-
quality academic work.
Lee et al 2023 USA Review Peer- Benefits, Limits, and The use of artificial intelligence (AI) in medicine is expanding, including applications
article reviewed Risks of GPT-4 as an AI in medical image analysis, drug interaction detection, patient risk identification, and
Chatbot for Medicine medical note coding.
This article focuses on a specific type of AI called a medical AI chatbot, utilizing the
GPT-4 system developed by OpenAI.
A chatbot consists of a general-purpose AI system and a chat interface, allowing
users to enter queries in natural language and receive relevant responses.
Prompt engineering is essential for optimizing chatbot performance, as current
systems are sensitive to the wording and form of prompts.
While GPT-4 excels in providing correct answers to questions with firm answers, its
most interesting interactions occur when prompts have no single correct response.
Chatbot errors, referred to as hallucinations, can be dangerous in medical
scenarios, so it is crucial to verify the output.
GPT-4 can help catch mistakes, not only in its own work but also in human-
generated content, providing a means of error detection and correction. GPT-4 is a
general-purpose AI system developed by OpenAI, designed to have broad
cognitive skills rather than being programmed for specific tasks in medicine.
Prompts for GPT-4 can be in the form of questions or directives to perform specific
tasks, and they can be written in various human languages and include data inputs
like spreadsheets, research papers, and equations.
Microsoft Research and OpenAI have been studying the applications of GPT-4 in
healthcare and medicine, focusing on areas such as documentation, data
interoperability, diagnosis, research, and education.
Other notable AI chatbots like LaMDA (Google) and GPT-3.5 have also been
explored for medical applications, despite not being explicitly trained for healthcare
purposes.
35
reviewed models in medicine and such as ChatGPT, because they may be utilised for text production,
medical research summarization, and correction activities.
LLMs can produce content for patient-friendly communication, standardised
reporting, presentation outlines, sample cover letters, and clinical documentation.
As LLMs mirror the training data, which may perpetuate prejudices and
underrepresent some perspectives, ethical problems include the possibility of
hidden biases in LLMs.
Trust and credibility of LLM-generated content are challenges, as LLM outputs are
difficult to trace and may be inaccurate. Current models do not evaluate quality or
provide measures of uncertainty.
Authorship becomes a complex issue when LLMs are used, as LLMs lack
intentionality and cannot approve or be held accountable for the final work. New
guidelines may be required.
Varying payment models for LLMs may widen digital divides, and affordability and
accessibility should be ensured.
Privacy concerns arise regarding the collection, use, and potential dissemination of
data inputted into LLMs. Strict controls and informed consent are necessary.
It is important to establish guidelines that promote responsible and effective use of
LLMs, rather than imposing an outright ban on their use.
In summary, while LLMs have the potential to revolutionize medicine and medical
research, careful consideration of their biases, trustworthiness, authorship,
equitability, and privacy implications is crucial. Responsible guidelines should be
established to ensure their ethical use.
Loh 2023 Australia Review Peer- ChatGPT and ChatGPT, an AI chatbot tool released by OpenAI, has gained significant popularity
article reviewed generative AI chatbots: and raised concerns about fraud and plagiarism due to its ability to generate text
challenges and similar to human-authored content.
opportunities for The risks of generative AI tools in health include challenges of fraud and plagiarism
science, medicine and in educational settings, the potential for biased outputs, and privacy and
medical leaders cybersecurity concerns.
AI tools like ChatGPT have been used to generate scientific articles, leading to
changes in guidelines by scientific meetings and publishers to address authorship
and credibility.
Despite the risks, generative AI tools offer opportunities for scientific research,
including assistance in experimental design, peer review, and administrative tasks
such as grant applications and research article editing.
In clinical practice, AI chatbots like ChatGPT can support frontline clinicians by
automating administrative tasks, such as generating discharge summaries and
medical letters.
AI chatbots can also assist in patient communication by providing accurate
information and answering low-risk health questions, although ethical
considerations need to be addressed.
Some health systems have already started using generative AI technology to
improve healthcare provision and alleviate administrative burdens and clinician
37
burnout.
Li 2023 China Commentary Peer ChatGPT has made the ChatGPT is a language generation model developed by OpenAI and is capable of
reviewed field of surgery full of generating high-quality text content.
opportunities and It has been trained on a large amount of preprocessed and filtered text data.
challenges IBM's Watson robot is an AI technology specifically designed for the medical field,
providing diagnosis and treatment suggestions based on clinical guidelines.
In the field of surgery, the Da Vinci robotic surgical system is widely used, but
surgeons still operate the surgeries and AI technology assists in surgical planning,
navigation, and support.
While ChatGPT can provide valuable diagnostic and treatment suggestions,
doctors must evaluate and judge these suggestions and comply with legal
regulations.
ChatGPT is suitable for document polishing and has the potential to be a powerful
medical assistant in the future.
The current application of ChatGPT in the medical field may not have a disruptive
effect yet, but its potential is promising and it can become a valuable tool with
further development and training.
Liebrenz et al UK Commentary Peer Generating scholarly The use of generative AI, such as ChatGPT, in medical publishing practices has
2023 reviewed content with ChatGPT: unknown implications, but it raises substantial ethical concerns.
ethical challenges for ChatGPT is an AI chatbot developed by OpenAI that generates responses based
medical publishing on internet sources and has been used to create university essays and scholarly
articles.
The imperceptibility of AI-generated content to human readers and anti-plagiarism
software raises concerns about copyright, attribution, plagiarism, and authorship.
There is a debate about whether AI can fulfill the criteria for authorship set by
organizations like the International Committee of Medical Journal Editors.
Organisations like the Committee on Publication Ethics and the International
Association of Scientific, Technical, and Medical Publishers have given ethical
principles and guidelines for AI in publishing.
As AI technologies proliferate, comprehensive talks about authorship laws and AI-
generated work are seen as urgent and crucial.
The usage of ChatGPT and comparable AI platforms might result in a paywall-
based business model, escalating already-existing inequities in scientific
publication. ChatGPT's accessibility and ease of use have the potential to increase
scholarly output and democratize knowledge dissemination in multiple languages.
However, there are concerns about the potential harm caused by ChatGPT in
producing misleading or inaccurate content, contributing to scholarly
misinformation.
Competitors to ChatGPT may arise, amplifying the challenges associated with AI-
generated content in scholarly publishing.
The Lancet Digital Health and the Lancet family are called upon to carefully
consider the ethical implications of publishing articles produced by AI and initiate
discussions on comprehensive guidance for AI-generated content in scholarly
38
publishing.
Lin 2023 Hong Kong Commentary Preprint Modernizing authorship The conventional requirement for authorship, based on "substantial contributions,"
criteria: Challenges is out of sync with current practices and emerging trends in scientific research.
from exponential The exponential increase in the number of authors per paper poses challenges to
authorship inflation and the notion of "substantial contributions," as each author's contribution becomes
generative artificial increasingly small.
intelligence The rise of generative AI tools, like ChatGPT, in scientific research raises questions
about whether they should be listed as co-authors when they substantially
contribute to a paper.
The International Committee of Medical Journal Editors (ICMJE) and other
authoritative guidelines emphasize substantial contributions to the research and
writing process as criteria for authorship.
However, the current criteria do not align with the increasing number of authors and
the potential substantial contributions of AI tools.
A new approach is proposed, where authorship is seen as proportional, role-
specific credit, reflecting the specific contributions made by individuals in various
roles.
The CRediT system, which defines specific roles in research, can be used to
specify and quantify individual contributions in a standardized manner.
Organizations, publishers, and authors should consider reevaluating and updating
authorship criteria to better reflect current research practices and the involvement
of generative AI tools.
Recommendations include revising the requirement of "substantial contributions,"
requiring authors to specify contributions using a standardized system like CRediT,
and actively documenting and discussing contributions throughout the research
process.
Maeker and France Commentary Peer- ChatGPT: A solution for Artificial intelligence (AI) can produce grammatically correct and coherent text
Maeker-Poquet reviewed producing medical based on patterns it has learned, but it lacks a deep understanding of the content.
2023 literature reviews?
AI like ChatGPT operates by processing natural language and following patterns in
data, not by truly understanding the content.
The initial trial showed that while ChatGPT could provide clear, well-written content,
it lacked sufficient references, often invented content, and did not meet established
literature review guidelines (e.g., PRISMA and ICMJE).
A second trial, with more structured input and guidance, led to more coherent
39
content but still had shortcomings in terms of depth, reference handling, and
staying on topic.
Despite the limitations, ChatGPT could be a useful tool in drafting early versions of
literature reviews or helping novice writers get started, owing to its speed and
ability to generate content.
AI could also help suggest initial manuscript structure, rephrase and summarize
content, and suggest references, although it can only draw from information up to
2021 and can't compare and interpret results.
Guidelines set by publishers like Elsevier allow for the use of AI for improving
readability and language but discourage it from replacing key research steps like
producing scientific information, analyzing data, or formulating conclusions.
The use of AI such as ChatGPT in writing scientific literature has potential, but
should be considered a supplement to, not a replacement for, human expertise and
judgement. Ethical issues around this usage remain under debate.
Marchandot et al France Commentary Peer- ChatGPT: the next ChatGPT, developed by OpenAI, is a powerful language model capable of
2023 reviewed frontier in academic generating human-like text. It can understand context and generate coherent, fluent
writing text that is often indistinguishable from text written by humans.
for cardiologists or a
pandora’s box of ethical The model has potential applications in academic research, such as assisting in
dilemmas data analysis, literature reviews, and drafting research papers. It can analyze vast
amounts of data quickly and can be trained on specific topics to help find and
summarize relevant literature.
There is ongoing debate on the use of AI tools like ChatGPT in academic research.
40
Real-time monitoring of ECG signals from wearable electronic devices and cardiac
implants can be facilitated by AI, alerting medical professionals or patients when
significant changes occur based on timing, duration, and context.
By reducing noise, artifacts, and interference, as well as extracting data that may
not be visible to the human eye, such as heart rate variability, beat-to-beat
intervals, and wavelet transforms, AI can enhance signal processing, leading to
improved ECG quality and accuracy.
The fusion of ECG data with information from other modalities, including imaging,
genomics, proteomics, and biomarkers, can be facilitated by AI.
Mehnen et al Austria Original Preprint ChatGPT as a medical The study tested the diagnostic accuracy of ChatGPT using 50 clinical case
2023 article doctor? A diagnostic vignettes, including 40 common complaints and 10 rare diseases.
accuracy study on
41
common and rare The case vignettes were collected from previous research and from the first
diseases matching articles on PubMed for randomly selected rare disorders with an
approved orphan drug.
ChatGPT was prompted with the phrase “What are the 10 most likely diagnoses for
this patient?” followed by the full text of the clinical case vignette.
Each vignette was prompted three times in independent chats using both ChatGPT
3.5 and ChatGPT 4, totaling 300 prompts and 3000 suggested diagnoses.
The 50 case vignettes were also presented to three human medical doctors for
diagnosis without the use of search engines.
For common cases, ChatGPT 3.5 solved more than 90% of all cases within the top
two suggested diagnoses, while ChatGPT 4 solved 100% of all cases within the top
three suggested diagnoses.
For rare cases, ChatGPT 3.5 achieved an accuracy of 60% within the top 10
suggestions, while ChatGPT 4 achieved an accuracy of 90% within eight or more
suggestions.
The legal position of physicians using these models is unclear. However, the
precedent of clinical practice guidelines may provide some guidance. In cases
where physicians have followed these guidelines, it has often constituted
exculpatory evidence in malpractice suits.
LLMs have some distinctive features that make them different from clinical practice
guidelines. They have a tendency to generate incorrect outputs (hallucinations),
42
often without clear references for physicians to assess the reliability of the
information.
Most LLMs are trained on a broad array of web text, which does not discern
between reputable medical articles and less reliable sources such as online
forums. Efforts are being made to train LLMs on authoritative medical texts, but
these are still in their infancy.
LLMs' outputs are not fixed, they can yield different responses for the same query,
and also vary depending on the date and the phrasing of the question. This can
make it difficult to determine the reasonableness of a physician's actions based on
the LLM's output.
While clinical decision support tools often undergo rigorous validation processes,
the same can't always be said for LLMs. They are mainly validated as text
generators and their design and evaluation is often performed by computer
scientists, not clinicians.
Despite these drawbacks, LLMs have their benefits. They can provide more
patient-specific recommendations than other decision-support tools and may also
suggest diagnostic and treatment possibilities that may be overlooked by
physicians.
There are studies evaluating LLMs' performance in clinical scenarios, and although
the results show some accuracy, there are also instances where they provide
completely incorrect answers.
At the present time, physicians should only use LLMs to supplement more
traditional forms of information seeking, such as searching for reputable sources on
Google or using clinical decision support systems.
Future reliable LLMs may likely come from specialized systems rather than
generalist systems like ChatGPT. These specialized systems can be validated
more easily and are less prone to the problems of indiscriminate information
sourcing.
The writers of the article propose a precise description of the registry's objective,
target market, and domain, as well as frequent evaluations for preserving its worth
and applicability.
In order to raise awareness of the register among members of the Royal College of
Radiologists (RCR) and other interested parties, ChatGPT can also provide
precise, coherent, and correct text.
The AI model's comprehension of English may also help with the registry's long-
term upkeep by detecting pertinent updates, summarising fresh research, and
creating rules to make sure the registry is kept current.
The incorporation of ChatGPT in the creation and upkeep of the registry may
simplify monitoring, assessment, and communication procedures, potentially
enhancing patient safety, resource effectiveness, and learning retention.
The introduction of ChatGPT into the radiology AI deployment registry has the
potential to make a substantial impact on the development of AI technology.
Nastasi et al 2023 USA Commentary Preprint Does ChatGPT Provide A huge language model from OpenAI called ChatGPT has demonstrated amazing
Appropriate and ability in conversational activities like successfully choosing the right radiological
Equitable Medical studies, writing research papers, and passing medical licencing examinations.
Advice?: A Vignette-
Based, Clinical However, due to a lack of publicly available training data and unsolved concerns
Evaluation Across Care regarding the safety, fairness, and regulation of AI systems, there are doubts about
Contexts its capacity to offer competent and equitable medical advice in a variety of clinical
scenarios.
It was discovered that ChatGPT alters its responses depending on the patient's
44
insurance status, race, and gender, frequently resulting in varied and clinically
unsuitable recommendations that, if followed, could worsen health inequities.
The study's limitations include the fact that just three distinct clinical settings were
tested, and that within-vignette variance was not measured.
Despite these drawbacks, the study offers crucial data that contextualises
ChatGPT's capacity to deliver accurate and impartial guidance along the treatment
continuum.
The results showed that ChatGPT's diagnostic accuracy was higher than that of the
general medical doctors but lower than that of the neurologist specialists. However,
when considering the three most likely diagnoses or the five most likely diagnoses
provided by ChatGPT, the accuracy reached a level comparable to the experts.
The study also found that ChatGPT performed equally well in diagnosing acute
neurological disorders compared to human doctors, but its accuracy decreased for
non-acute cases.
The study acknowledges the ethical, integrity, and data safety concerns associated
with using AI in medicine. While AI has the potential to improve healthcare, careful
consideration of these concerns is necessary.
North UK Commentary Peer Plagiarism re-imagined The author of this personal account reflects on a time when scientific writing was
2023 reviewed done manually, with quills, pens, and typewriters. Manuscripts were often typed by
someone else, and research was conducted using limited resources like tape
recorders and index cards. The author received a request to review a manuscript
from the journal Life Sciences, which turned out to have significant portions copied
verbatim from a paper the author had previously published. The author contacted
the editor, who took the matter seriously and barred the authors from submitting
further work to the journal.
Later, the author came across another paper published in Neuroscience Letters,
which also contained substantial portions copied from the author's earlier work. The
46
author contacted the editor, who published a disclaimer acknowledging the overlap
and the prior publication of the author's work. The author reflects on the irony of
receiving a plagiarized manuscript for review and the increased ease of plagiarism
in the modern age with AI-based text generators like ChatGPT. The author also
highlights the detrimental impact of simplistic numerical methods, such as the h-
index, on scientific integrity and the motivation behind academic dishonesty.
Oh et al 2023 Korea Original Peer ChatGPT goes to the The goal of this study was to assess how well the ChatGPT language model
article reviewed operating room: performed using a dataset of questions from the Korean general surgery board
evaluating GPT-4 exam. 280 questions from the exam's first stage between 2020 and 2022 made up
performance and its the dataset. The dataset did not include any questions that requested visual data.
potential in surgical For model testing, the study used ChatGPT versions GPT-3.5 and GPT-4.
education and
training in the era of The study findings revealed a notable disparity in performance between the GPT-
large language models 3.5 and GPT-4 models, with GPT-4 exhibiting significant improvement. GPT-4
achieved an overall accuracy rate of 76.4%, surpassing the accuracy rate of GPT-
3.5, which was 46.8%. The accuracy rates varied across different subspecialties,
with transplantation, breast, and hepatobiliary and pancreas demonstrating higher
accuracy rates.
These results suggest that language models can have implications for surgical
education and training, providing a tool for evaluating surgical knowledge and skills.
However, further research and fine-tuning of the models are needed to enhance
their performance and applicability in the surgical field.
Okan 2023 Turkey Commentary Peer AI and Psychiatry: The Modern language model ChatGPT, created by OpenAI, simulates human
reviewed ChatGPT Perspective conversation and offers precise information and responses on a range of subjects.
ChatGPT can support the selection of treatments and long-term patient monitoring
in the field of psychiatry. It can also help with the diagnostic process, support
psychiatric education, simulate case examples, assist researchers in data
collecting and analysis, and more.
These technologies can help psychiatrists make more accurate diagnoses, design
effective personalized treatments, improve clinical trial accuracy and speed, and
increase accessibility to mental health care through chatbots and virtual assistants.
The use of AI in psychiatry has the potential to advance the field and improve
patient outcomes.
The integration of AI in India's healthcare system holds the potential to bring about
transformative changes in healthcare delivery. However, it is crucial to exercise
caution and consider all relevant factors to ensure a well-planned and regulated
implementation process.
Ray and India Letter to Peer- AI Tackles Pandemics: The research examines ChatGPT's possible uses as an AI language model for
Majumder 2023 editor reviewed ChatGPT's Game- infectious disease management and control.
Changing Impact on ChatGPT should be regarded as an additional tool for clinical practise rather than a
Infectious Disease replacement for healthcare providers' knowledge.
Control The problems of ChatGPT's accessibility, lack of human interaction, and inaccurate
information are all acknowledged.
Powered by deep learning artificial intelligence (AI) techniques and based on the
GPT-3 architecture, ChatGPT provides human-like responses to natural language
queries.
ChatGPT successfully answered all questions related to the case, offering well-
explained reasoning. Both the initial response and a regenerated response
provided satisfactory results.
There is potential for future iterations of ChatGPT, specifically adapted for medical
use, to be valuable in handling less common clinical cases that experts may
sometimes overlook.
The authors envision a future where AI serves as a tool utilized by clinicians, rather
than a replacement for them. They predict that in the coming years, clinicians who
incorporate AI into their practice will replace those who do not.
Sallam Jordan Review Preprint The Utility of ChatGPT A thorough analysis was done to determine the possible benefits and drawbacks of
2023 as an Example of Large the "ChatGPT" AI-based large language model (LLM) in healthcare instruction,
Language Models in study, and practise.
Healthcare Education,
Research and Practice: The review employed PRIMSA standards and utilised the term "ChatGPT" to
Systematic Review on search English records in PubMed/MEDLINE and Google Scholar. Any preprints or
the Future Perspectives published studies considering ChatGPT in the context of healthcare were eligible
51
ChatGPT was mentioned in 85% of the records that were collected, with its use in
scientific writing and medical research being the most common. It was praised for
its ability to reduce costs, facilitate paperwork, enable personalised therapy, and
promote more health literacy in healthcare practise.
ChatGPT could be a useful resource for individualised learning in the field of health
education and promote a move towards critical thinking and problem-based
learning.
ChatGPT does not meet the criteria to be an author in scientific journals as of right
now. However, modifications to the ICMJE/COPE standards may cause this to
change.
The study urges the creation of a code of ethics and conduct for the responsible
use of ChatGPT and other LLMs, engaging all parties involved in healthcare
education, research, and practise.
Sallam et al Jordan Review Peer ChatGPT Output ChatGPT could be a useful resource for individualised learning in the field of health
2023 reviewed Regarding Compulsory education and promote a move towards critical thinking and problem-based
Vaccination and learning.
COVID-19 Vaccine
Conspiracy: A In light of its impending integration into web search engines, the study sought to
Descriptive Study at the assess the output of the AI-based language model ChatGPT in relation to COVID-
Outset of a Paradigm 19 vaccine conspiracy theories and mandatory vaccination.
52
Findings revealed that ChatGPT replies rejected conspiracy theories regarding the
origins of the SARS-CoV-2 virus as untrue and devoid of scientific support.
Additionally, ChatGPT answers substantially refuted the claims of COVID-19
vaccine conspiracies.
Salvagno et al Italy Review Peer Can artificial ChatGPT has demonstrated its usefulness in supporting researchers across
2023 article reviewed intelligence help for various aspects of scientific writing, including drafting articles, abstracts, literature
scientific writing? research, data summarization, structure suggestions, and language review.
However, there have been no published articles so far on the use of this technology
in critical care medicine, leaving it uncertain whether its implementation would
simplify or complicate the writing process within this discipline.
While ChatGPT cannot generate entirely new ideas, it can assist researchers by
developing their existing ideas and creating initial drafts. It has the potential to aid
in the review process by identifying and summarizing academic papers and
highlighting areas of uncertainty.
During the writing process, ChatGPT can generate an initial draft of a scientific
paper, suggest titles, justify sample sizes, and describe data analysis techniques.
Despite its potential benefits, the technology's limitations must be recognized. The
accuracy, current knowledge, and precision of the AI-generated content may be
variable and should always be validated by human experts.
Ultimately, AI and chatbots like ChatGPT should not replace human researchers'
expertise, judgment, and personality, but rather supplement them.
Sanmarchi et al Italy Original Preprint A step-by-step Aim of the Study: The purpose of this research was to evaluate how early-stage AI-
2023 article Researcher's Guide to based transformer models, specifically OpenAI's ChatGPT, can assist researchers
the use of an AI-based in the multiple steps of an epidemiological study. The study was simulated using
transformer in the STROBE framework and the responses of the transformer were assessed for
epidemiology: an coherence and relevance.
exploratory analysis of
ChatGPT using the Methods: The study involved posing questions derived from the STROBE checklist
STROBE checklist for to ChatGPT and assessing its answers in the context of conducting an
observational studies epidemiological study. The answers were evaluated by three independent senior
researchers.
54
Expert Assessment: The outputs of the transformer were assessed through "human
evaluation," where annotators were asked to rate the responses produced by the
model on a 1 to 5 points Likert scale, taking into account the coherence and
relevance of the responses.
In the study, the output of ChatGPT was compared to the diagnosis or treatment
recommendations made by the ID or CM specialists in the identical clinical settings.
On a scale of 1 (bad, wrong advise) to 5 (outstanding, entirely correlating with the
specialist's recommendations), the AI's output was evaluated.
The ChatGPT had a mean score of 2.8 and a standard deviation of 1.5. It
performed best when interpreting positive blood culture results (mean 3.3, SD 1.3)
and poorly when providing guidance on osteomyelitis or infections of prosthetic
joints (mean 1.3, SD 0.5). These scores' differences were statistically significant.
ChatGPT occasionally supplied unclear treatment options and didn't always take
into account given data. It occasionally made incorrect claims about the origin of its
guidance, which raises questions about its veracity and accuracy.
55
The study's limitations include the exclusion of chronic cases and the possibility of
differences between the data provided to ChatGPT and the reference standard
clinical consultation.
The study comes to the conclusion that while ChatGPT can generate moderately
accurate diagnostic or therapeutic suggestions, it cannot take the role of clinical
expertise. Clinicians are required to direct and interpret ChatGPT's
recommendations.
Schulte 2023 USA Original Peer Capacity of ChatGPT to This study aimed to evaluate the capability of ChatGPT in suggesting suitable
article reviewed Identify Guideline- systemic treatments for new cases of advanced solid malignancies, in comparison
Based Treatments for with the recommendations found in the National Comprehensive Cancer Network
Advanced Solid Tumors (NCCN) guidelines.
The study used 51 different diagnoses and ChatGPT was able to identify 91
different medications for the treatment of advanced solid tumors.
The overall VTQ was found to be 0.77, indicating a high level of concordance with
the NCCN guidelines.
For each case, ChatGPT was able to provide at least one systemic therapy
example that matched the NCCN's suggestions.
There was a weak correlation identified between the incidence of each malignancy
and the VTQ.
The study concluded that ChatGPT showed significant ability in recognizing and
suggesting medications used in the treatment of advanced solid tumors according
to the NCCN guidelines.
The study anticipates that future versions of ChatGPT will have improved accuracy
and consistency in this area and emphasizes the need for further studies to
quantify its capabilities more effectively.
56
Singh 2023 India Commentary Peer Artificial intelligence in OpenAI launched ChatGPT, an AI-based chatbot, on November 30, 2022, and it
reviewed the era of ChatGPT - has already amassed over 100 million users. Numerous other AI applications are
Opportunities and on the horizon, including Google Bard, Microsoft Bing AI, Chinese Ernie bot,
challenges in mental Korean SearchGPT, Russian YaLM 2.0, Chatsonic, Jasper Chat, Character AI,
health care Perplexity AI, and YouChat.
AI platforms like ChatGPT are recognized for their vast potential across various
domains, including mental health. They are being utilized for chatting, gaming,
computer programming, and even generating scientific papers and abstracts.
However, while AI chatbots like ChatGPT have immense promise, there are
concerns regarding their use in mental health. These concerns include the potential
for providing incorrect information or advice, issues related to confidentiality,
privacy, and data security, as well as challenges in accurately diagnosing mental
health conditions.
Operative Notes intraocular injections, open-globe injury repair, extraocular muscle surgery, Levator
resection for ptosis, and dacryocystorhinostomy. The AI model's outputs were
evaluated for their evidence-based content, specificity, factual accuracy, and ability
to recognize mistakes.
The AI model could quickly generate detailed discharge summaries and operative
notes in response to provided prompts, with responses usually returned in less
than 20 seconds.
While the AI model's summaries were generally accurate, they contained some
generic text, particularly in the 'discharge instructions' section.
The operative notes were divided into several sections, including pre- and post-
operative diagnosis, procedure, anesthesia, indication, procedure details, post-
operative care, estimated blood loss, complications, and summary.
Despite some shortcomings, the authors noted that the model's responses could be
customized to the user's needs with further training. This study demonstrates the
potential of AI in assisting with documentation tasks in the field of ophthalmology.
Tang et al 2023 USA Review Preprint Evaluating Large This extensive study assesses the performance of large language models (LLMs)
Language Models on in medical evidence summarization using Cochrane Reviews across six clinical
Medical Evidence domains, including heart failure, kidney illness, esophageal cancer, Alzheimer's
Summarization disease, and neurological problems.
The study compares the zero-shot performance of two models: GPT-3.5 and
ChatGPT. These models were given different parts of the review abstracts to
summarize. The models' output was assessed on several metrics including
ROUGE-L, METEOR, BLEU, and manual human evaluation for factors like
coherence, factual consistency, comprehensiveness, and harmfulness.
58
Findings indicate that all models perform similarly on automatic metrics with higher
ROUGE scores showing that key information from the source document was
effectively captured. However, the models were found to be more extractive
compared to human-written summaries, indicating a lower level of abstraction.
Human tests revealed that ChatGPT's summaries were more cohesive than
GPT3.5's. It is noteworthy that less than 10% of the summaries created by
ChatGPT-MainResult included factual errors. In terms of thoroughness, ChatGPT-
MainResult and ChatGPT-Abstract both delivered more than 75% of the time, with
the former also producing the fewest summaries that were medically dangerous.
Inclusion and Exclusion Criteria: Articles discussing ChatGPT within the context of
medical education, literature, or practice were included. Non-medical or unrelated
contexts, non-English sources, and articles published outside the specified date
range were excluded.
Results: A total of 65 papers were identified in PubMed and 110 papers in Europe
PMC. These papers highlighted the utilization of ChatGPT in diverse medical fields
such as medical education, scientific writing, research, and diagnostic decision-
making.
Key Findings: The review identified eight main themes: (1) medical writing, (2)
medical education, (3) diagnostic decision-making, (4) public health, (5) scientific
research, (6) ethical considerations of ChatGPT use, (7) ChatGPT's potential to
automate medical tasks, and (8) criticism of ChatGPT's usage.
ChatGPT in Public Health: ChatGPT can provide reliable estimates in public health
research, provided the input data is valid and accurate.
Despite its popularity, the AI has been noted to sometimes write plausible but
incorrect or nonsensical answers. This includes referencing non-existent scientific
studies.
Concerns have been raised about the impact of ChatGPT on education, as it can
write essays on a range of topics. While the program can generate factual answers,
the quality of scholarly writing is currently lacking.
Science journals are now updating their licenses and editorial policies to explicitly
state that text generated by AI tools, such as ChatGPT, cannot be used in their
work. This includes figures, images, or graphics produced by AI tools.
The journals recognize that most instances of scientific misconduct occur due to
lack of adequate human attention, hence the need for scientists to be vigilant in
their work.
The evaluation criteria were adapted from the EASE guidelines, focusing on three
major areas: structure, scientific content, and credibility.
Grammarly, another AI-based tool, was used to score two items: originality and
readability.
The team selected three diverse topics for human authors and ChatGPT to write
about, shared the output as blinded versions, and evaluators scored the articles
independently.
On scientific content, both types of articles scored well on originality, but ChatGPT
articles were weaker in specificity of response, lack of numerical data,
cohesiveness, and inclusion of study limitations.
Efficiency was the only area where ChatGPT scored perfectly, taking significantly
less time to produce the articles compared to human authors.
The study concluded that while ChatGPT can efficiently produce original and
seemingly coherent articles, they lack in-depth scientific basis, lack reproducibility,
and might quote non-existent scholarly work. This raises questions about the
accuracy and reliability of ChatGPT in the context of scientific writing.
Uprety et al 2023 USA Review Peer ChatGPT-A promising ChatGPT is a transformative AI model developed by OpenAI, offering
article reviewed generative AI tool and sophisticated, human-like text-based conversations. With the release of GPT-4,
its implications for ChatGPT Plus has seen widespread interest and potential applications, particularly
cancer care within the healthcare sector.
ChatGPT can be utilized in the medical field for extracting crucial information from
patient records, saving physicians time and enhancing efficiency. This is
particularly beneficial in areas such as oncology, where patients often have
extensive and complicated histories.
The AI can be used for administrative tasks, such as creating insurance letters for
evidence-based authorization of therapies.
ChatGPT could aid in the rapidly evolving field of oncology by staying updated on
recent advances and approvals, aiding in the interpretation of complex Next-
Generation Sequencing (NGS) reports, and making appropriate treatment
recommendations based on the correlation between mutations (biomarkers) and
treatment drugs.
The AI can support oncologists in making suitable clinical trial recommendations for
patients by interpreting their detailed medical history and the range of content
available on ClinicalTrials.gov.
Despite its vast potential, ChatGPT currently has some limitations. It's not
compliant with the Health Insurance Portability and Accountability Act (HIPAA),
posing a risk of violating patient privacy. Additionally, the AI could unintentionally
propagate biases and misinformation inherent in the internet text it was trained on,
and it is vulnerable to adversaries and manipulation.
The authors anticipate that future iterations of ChatGPT may overcome these
limitations and become a highly valuable tool for healthcare, particularly in the field
of oncology.
Uz and Umay Turkey. Original Peer "Dr ChatGPT": Is it a The objective of this study was to utilize Google Trends to identify the most
2023 article reviewed reliable and useful frequently searched terms related to common rheumatic disorders (OA, RA, FMS,
source for common AS, SLE, gout, and PSA). Based on the search engine results, keywords were
rheumatic diseases? determined and categorized for each condition. Likert-type ratings were employed
to evaluate the accuracy and value of ChatGPT's responses to these keywords.
Using Likert-type scales, the study evaluated the accuracy and value of ChatGPT's
responses for each keyword. The reliability and usefulness scores demonstrated
good to very good agreement, as indicated by Cronbach values and inter-rater
reliability scores.
Van Dis et al 2023 the Commentary Peer- ChatGPT: five priorities Large Language Models (LLMs) like ChatGPT have the potential to revolutionize
Netherlands reviewed for research the scientific research field, offering assistance in various tasks including writing
essays, summarizing literature, and identifying research gaps.
However, there is controversy around this technology due to its capacity to produce
convincingly wrong information, potentially distorting scientific facts and spreading
misinformation.
LLMs can introduce inaccuracies, bias, and plagiarism. ChatGPT, for instance, has
been found to generate false and misleading responses when asked complex,
nuanced scientific questions.
Researchers using AI like ChatGPT risk being misled by false or biased information
and incorporating it into their work. Unintentional plagiarism is also a concern as AI
reproduces text without reliable citations.
There are emerging questions about authorship and rights to texts created with AI
assistance, indicating the need for updated definitions and legal considerations.
AI could change the academic skill set, optimizing training, reducing the need for
certain skills, and introducing new ones. As AI advances, it might handle more
complex tasks, prompting the need for careful evaluation of AI acceleration vs loss
of human potential.
64
Every research group should discuss the use and implications of LLMs, and
educators should introduce discussions about LLM usage and ethics to students.
Accountability for research work will apply regardless of whether it's generated with
ChatGPT or not.
Waisberg et al Ireland Letter to the Peer GPT4: a new era of A sizable language model called ChatGPT was created by OpenAI. To anticipate
2023 editor reviewed artificial intelligence in words or phrases in context, it employs a deep neural network based on the
medicine Transformer architecture.
The AI processes user input, analyses it, and then responds in accordance with the
patterns it identified during training. This response is intended to be situationally
suitable and can cover a wide range of issues.
Banking, gaming, and healthcare are just a few of the industries where artificial
intelligence is being used. It might help with patient diagnosis, treatment
alternatives, and individualised care plans, for instance, in the healthcare industry.
The AI was instructed to write an easy-to-read discharge narrative for a patient who
had undergone a simple cholecystectomy. Important recommendations on exercise
limitations, dietary modifications, and when to seek medical assistance were all
included in the summary.
The AI was also asked to provide information about the latest clinical trials to treat
interstitial lung disease (ILD). It listed four ongoing phase 2 clinical trials and
provided brief details about each one.
Furthermore, the AI was asked to provide the latest guidelines for AI in medicine. It
listed guidelines from the European Commission, the American Medical
Association, the International Medical Informatics Association, and the National
Institute for Health and Care Excellence, all of which were relevant.
65
In contrast, the AI's newly implemented image analysis feature demonstrated a lack
of accuracy in identifying fundus photographs of anterior ischemic optic neuropathy
and non-arteritic anterior ischemic optic neuropathy. Instead, it erroneously
classified these images as a schematic representation of the CRISPR-Cas system.
ChatGPT as an Aide for Scientific Innovation: Despite its limitations, ChatGPT can
be used beneficially in scientific research to improve completed research papers. It
can increase researchers' productivity, save time, and improve the quality of their
content.
Pros and Cons of ChatGPT in Scientific Research and Publishing: While ChatGPT
can be useful for certain tasks like editing, it is not reliable for providing accurate
facts or references. Using it judiciously while being aware of its limitations can help
researchers streamline their work without risking the publication of false
information.
Xue et al 2023 China Commentary Peer- The potential impact of ChatGPT is an AI model developed by OpenAI that simulates human interaction
reviewed ChatGPT in clinical and and uses a deep learning technique called 'transformer architecture'. It's trained on
translational medicine large datasets of text from the internet.
It is still unclear how ChatGPT will affect clinical medicine. However, similar AI
models could help with patient education, clinical trial recruiting, clinical data
management, research support, clinical decision support, and the automation of
activities like data processing and picture interpretation.
AI may have unfavourable effects on society, such as privacy problems and bias or
discrimination. To ensure the responsible and effective use of AI in healthcare,
more research and development is required.
Even so, ChatGPT and other forms of AI will continue to advance. The best course
of action is to embrace it, making use of its potential to advance clinical practise
while simultaneously addressing any potential drawbacks.
Yadava 2023 India Editorial Peer- ChatGPT—a foe or an ChatGPT, an AI platform developed by OpenAI, employs natural language
reviewed ally? processing technology, using advanced machine learning paradigms such as
transfer learning, supervised learning, and reinforced learning.
67
The model has the capability to generate articulate and meaningful content across
various fields of knowledge, with outputs including music, plays, poetry, and song
lyrics.
Despite its advanced abilities, there are concerns within academia about the impact
of non-human authors on the integrity and validity of scientific publications. This
could potentially lead to an increase in academic plagiarism.
Some renowned scientific journals such as 'Nature' and 'JAMA Network Science'
have decided not to accept articles generated by ChatGPT, demanding full
disclosure of its use.
The technology has also sparked worries among non-medical intelligentsia and
politicians, with fears about potential job losses, discrimination, and uncontrollable
military applications.
Historical examples show that disruptive technologies often face resistance upon
introduction. However, despite initial skepticism, many have gone on to have
profound impacts on society and industry.
Some academic journals and researchers have already adopted the technology,
recognizing its potential benefits for medical writing and document creation.
The debate on the implications of AI like ChatGPT is ongoing. It could turn out to
be a dual-use technology, where the ethics and intent behind its use determine its
value.
Regardless of the pros and cons, AI-powered language models are becoming a
reality, with human oversight remaining crucial.
Large Language Models (LLMs), like OpenAI’s GPT3 and Google’s PALM, Gopher,
and Chinchilla, have evolved to exhibit emergent properties: performing tasks they
68
were not explicitly trained on. This could be due to the models' ability to extract
more knowledge from vast amounts of text data used in their training.
Current LLMs face challenges such as mirroring biases and inaccuracies from their
training data, raising concerns about their suitability for critical fields like healthcare.
Moreover, LLMs trained on biomedical corpora, while being domain-specific, don't
necessarily reflect the realities of actual patients and diseases in healthcare.
Some LLMs, such as GatorTron and BEHRT, are trained on de-identified clinical
notes or disease classification codes to circumvent the problem of sensitive patient
data.
LLMs can exhibit biases and associations based on their training text, leading to
potential issues of racial-ethnic disparities in treatment recommendations.
LLMs can also generate "hallucinations" or false information, especially when faced
with insufficient or misleading information in the prompt.
Although LLMs have made encouraging strides in natural language processing and
creation, it is still unclear whether they are ready to be used as clinical tools that
patients will see. Focus should be placed on domain-specific training data, expert
clinician fine-tuning, and transparent depiction of output relevancy versus safety
implications in order to use AI chatbots in healthcare in a safe manner.
Young et al 2023 USA Research Peer The utility of ChatGPT The study aimed to assess the effectiveness of OpenAI's ChatGPT, a large
Letter reviewed in generating patient- language model (LLM), as an informational platform for patients seeking
facing and clinical information about melanoma care. To evaluate its suitability, twenty-five
responses for hypothetical patient questions related to melanoma were presented to three board-
melanoma certified dermatologists. The dermatologists assessed the appropriateness of the
69
However, the study found that only 64% of the responses were deemed sufficient
for clinical practice. This was mainly due to important details being omitted in
ChatGPT's advice. For instance, while recommending "regular skin exams" is
accurate, it lacks the specificity of a physician's recommendation for exams every
few months.
While ChatGPT generates responses that are well-written and sound plausible, it
has been observed that these responses often contain significant errors and
fabricated information. Consequently, ChatGPT cannot be considered a reliable
and trustworthy source for scientific writing.
Given the potential for falsifications and fabrications in ChatGPT's output, the study
70
concludes that there is no clear advantage to using ChatGPT for scientific writing.
Authors utilizing ChatGPT must take the responsibility to manually verify all facts,
statements, and references generated by the AI.
It is important to note that the authors listed on an article are ultimately responsible
for ensuring accuracy and integrity, not ChatGPT itself.
Detecting fabrication or falsification during the peer-review process of manuscripts
containing text generated by ChatGPT is challenging. Studies have shown that
human reviewers can overlook up to 32% of fully fabricated abstracts produced by
ChatGPT.
Directly adopting full text written by ChatGPT may constitute plagiarism and violate
the code of conduct in scientific publishing.
Scientific journals that accept articles involving ChatGPT in the writing process may
face a significant increase in retractions and loss of credibility.
While ChatGPT offers great potential for the future, its current state is not mature
enough for scientific writing. The role of a more advanced ChatGPT in scientific
writing necessitates comprehensive discussions and debates.
Considering the potential for fabricated and inaccurate information, the Science
family of journals has implemented a ban on all ChatGPT-generated content. It is
recommended that this policy becomes standard practice for all scientific
publishing.
Despite its limitations, ChatGPT can still be a useful tool for checking grammar and
syntax errors, as well as refining language, particularly for non-native speakers.
Zhong et al 2023 China Commentary Peer The Artificial Artificial intelligence (AI) can significantly enhance psychiatric research and
reviewed intelligence large practice by improving diagnostic accuracy, optimizing treatment outcomes, and
language models and offering personalized care. AI can analyze extensive patient data, recognize
neuropsychiatry complex patterns, and suggest individualized treatment strategies.
practice and research
ethic AI large language models (LLMs) like ChatGPT, despite their potential benefits,
raise concerns such as reliability and accuracy, transparency, accountability, and
ethical implications. These models can sometimes generate plausible yet incorrect
or nonsensical information, leading to doubts about their scientific reliability.
71
Accountability is another issue with AI LLMs. They could perpetuate biases present
in the training datasets, and this could lead to biased outputs. For instance, if the AI
model is trained on racially or ethnically biased data, its outputs may reflect these
biases.
Ethical concerns with AI in research include the potential for these models to
generate misleading content. Some cases have reported AI-generated abstracts
that fooled academic reviewers. AI LLMs' increasing usage for tasks like data
analysis, literature reviews, grant proposals, etc., is stirring debate about whether
they should be acknowledged as authors in research papers.
There are concerns over intellectual property rights violations as AI models are
trained on vast datasets that may include copyrighted or proprietary content.
There's a risk that these models could generate content that infringes on these
rights.
The article suggests measures to address these concerns and regulate AI LLMs
use in science. These include insisting on human review of AI-generated content,
establishing accountability mechanisms, investing in open-source LLMs, and
promoting a broad discussion among stakeholders on the implications and
challenges of AI.
In this particular case, ChatGPT produced a detailed medical report for a 31-year-
72
old male patient who presented with abdominal pain and sought medical attention.
The structure and organization of the case report were entirely generated by
ChatGPT, utilizing the patient's physical information and lab results as input.
The study aims to demonstrate that ChatGPT can consistently and accurately
produce coherent, comprehensive, and clinically relevant medical reports,
showcasing its potential in aiding healthcare professionals.
Zhu et al 2023 China Commentary Peer Can the ChatGPT and The researchers conducted an evaluation of multiple large language models
reviewed other Large Language (LLMs), including ChatGPT, YouChat, NeevaAI, Perplexity, and Chatsonic, to
Models with internet- assess their usefulness in providing accurate and comprehensive information about
connected database prostate cancer (PCa).
solve the questions and
concerns of patient with To evaluate the LLMs' performance, the researchers designed 22 questions based
prostate cancer? on patient education guidelines and their own clinical experience. They assessed
the LLMs in terms of accuracy, comprehensiveness, patient readability, humanistic
care, and stability.
Most of the LLMs exhibited accuracy rates above 90%, with ChatGPT
demonstrating the highest accuracy. Interestingly, the free version of ChatGPT
performed slightly better than the paid version.
One issue identified was the inclusion of outdated or incorrect information in some
LLM responses. Examples include inaccurate comparisons between apalutamide
and enzalutamide, and claims that open surgery was more common than robot-
73
However, the researchers caution against replacing doctors with LLMs at this
stage. LLMs may contain errors, omit important points, and struggle to analyze
specific contexts. They also lack the ability to ask follow-up questions for further
information or provide the same level of comfort as a human healthcare provider.
Zielinski et al 2023 UK Commentary Peer Chatbots, ChatGPT, Chatbots, including ChatGPT, are AI tools used in diverse fields, including
reviewed and Scholarly healthcare, customer service, and education. They can create new content by
Manuscripts: WAME processing and reorganizing existing information.
Recommendations on
ChatGPT and Chatbots Despite its potential, ChatGPT has several limitations. It can generate incorrect or
in Relation to Scholarly nonsensical answers and it may fail to ask clarifying questions. Its knowledge is
Publications restricted to what it learned before 2021.
Chatbots can only generate output based on their training data and they cannot
produce truly original thoughts. They may unintentionally plagiarize from their
training materials.
ChatGPT can produce false statements, though it does not possess the
intentionality to lie in the human sense.
As software tools, chatbots cannot be held legally responsible for their output.
Liability falls on the users.
AI tools like DALL-E 2 and Imagen, which generate images, have similar concerns
to ChatGPT. The provenance of images should be clearly indicated.
Editors need to establish policies on the use of AI tools like ChatGPT and require
tools for detecting AI-generated content.
NLP is already used in various applications like Siri, Alexa, Grammarly, and even in
automated journalism. In the medical field, NLP is used for tasks such as
information extraction, productivity improvement, and documentation.
ChatGPT, a product of OpenAI, has gained significant attention due to its ability to
generate human-like content, ranging from passing scores on the USMLE exam to
generating convincing medical abstracts.
While tools like Wordtune and Paperpal assist with sentence restructuring,
ChatGPT can help restructure entire manuscripts and provide feedback on
limitations.
ChatGPT is leading in its space, but it is not without competitors. Companies like
Google, Facebook, and Anthropic are working to create models with similar
capabilities.
Despite its impressive capabilities, ChatGPT has been met with criticism, including
the potential for generating confabulatory outputs (hallucinations) when it does not
have an appropriate response and potential bias.
Ethical considerations are also paramount when it comes to the originality of the
content produced by these tools, making transparency of use crucial.
Corporate investment in AI systems has been growing over the years, indicating a
growing interest and trust in the technology. However, the responsibility of using
75
321
322 10. Bauchner H. ChatGPT: Not An Author, But A Tool. Health Affairs Forefront. 2023. DOI: 10.1377/forefront.20230511.917632.
323
324 11. Baumgartner C. The potential impact of ChatGPT in clinical and translational medicine. Clin Transl Med. 2023 Mar;13(3):e1206. doi:
325 10.1002/ctm2.1206.
326
327 12. Benoit JR. ChatGPT for Clinical Vignette Generation, Revision, and Evaluation. medRxiv 2023.02.04.23285478; doi:
328 https://doi.org/10.1101/2023.02.04.23285478.
329
330 13. Bhattacharya K, Bhattacharya AS, Bhattacharya N, Yagnik VD, Garg P, Kumar S. ChatGPT in surgical practice—a New Kid on the Block.
331 Indian Journal of Surgery. 2023 Feb 22:1-4. https://doi.org/10.1007/s12262-023-03727-x.
332
333 14. Biswas S. ChatGPT and the Future of Medical Writing. Radiology. 2023 Apr;307(2):e223312. doi: 10.1148/radiol.223312.
334
335 15. Boßelmann CM, Leu C, Lal D. Are AI language models such as ChatGPT ready to improve the care of individuals with epilepsy? Epilepsia.
336 2023 May;64(5):1195-1199. doi: 10.1111/epi.17570.
337
338 16. Brainard J. Journals take up arms against AI-written text. Science. 2023 Feb 24;379(6634):740-741. doi: 10.1126/science.adh2762. Epub
339 2023 Feb 23.
340
341 17. Cahan P, Treutlein B. A conversation with ChatGPT on the role of computational systems biology in stem cell research. Stem Cell Reports.
342 2023 Jan 10;18(1):1-2. doi: 10.1016/j.stemcr.2022.12.009.
343
344 18. Nasrallah HA. A ‘guest editorial’… generated by ChatGPT? Current Psychiatry 2023:6. https://cdn.mdedge.com/files/s3fs-
345 public/CP02204006.pdf.
346
347 19. Cascella M, Montomoli J, Bellini V, Bignami E. Evaluating the Feasibility of ChatGPT in Healthcare: An Analysis of Multiple Clinical and
348 Research Scenarios. J Med Syst. 2023 Mar 4;47(1):33. doi: 10.1007/s10916-023-01925-4.
349
350 20. Chen S, Kann BH, Foote MB, Aerts HJ, Savova GK, Mak RH, Bitterman DS. The utility of ChatGPT for cancer treatment information
351 medRxiv 2023.03.16.23287316; doi: https://doi.org/10.1101/2023.03.16.23287316.
352
353 21. Cheng K, Wu H, Li C. ChatGPT/GPT-4: enabling a new era of surgical oncology. Int J Surg. 2023 May 16. doi:
354 10.1097/JS9.0000000000000451. Epub ahead of print.
355
77
356 22. Chervenak J, Lieman H, Blanco-Breindel M, Jindal S. The promise and peril of using a large language model to obtain clinical information:
357 ChatGPT performs strongly as a fertility counseling tool with limitations. Fertil Steril. 2023 May 20:S0015-0282(23)00522-8. doi:
358 10.1016/j.fertnstert.2023.05.151. Epub ahead of print.
359
360 23. Cifarelli CP, Sheehan JP. Large language model artificial intelligence: the current state and future of ChatGPT in neuro-oncology
361 publishing. J Neurooncol. 2023 May 20. doi: 10.1007/s11060-023-04336-0. Epub ahead of print.
362
363 24. Corsello A, Santangelo A. May Artificial Intelligence Influence Future Pediatric Research?-The Case of ChatGPT. Children (Basel). 2023
364 Apr 21;10(4):757. doi: 10.3390/children10040757.
365
366 25. D'Amico RS, White TG, Shah HA, Langer DJ. I Asked a ChatGPT to Write an Editorial About How We Can Incorporate Chatbots Into
367 Neurosurgical Research and Patient Care…. Neurosurgery. 2023 Apr 1;92(4):663-664. doi: 10.1227/neu.0000000000002414. Epub 2023
368 Feb 9.
369
370 26. Darkhabani M, Alrifaai MA, Elsalti A, Dvir YM, Mahroum N. ChatGPT and autoimmunity - A new weapon in the battlefield of knowledge.
371 Autoimmun Rev. 2023 May 19:103360. doi: 10.1016/j.autrev.2023.103360. Epub ahead of print.
372
373 27. Dave, M. Plagiarism software now able to detect students using ChatGPT. Br Dent J 234, 642 (2023). https://doi.org/10.1038/s41415-023-
374 5868-8.
375
376 28. Dave T, Athaluri SA, Singh S. ChatGPT in medicine: an overview of its applications, advantages, limitations, future prospects, and ethical
377 considerations. Front Artif Intell. 2023 May 4;6:1169595. doi: 10.3389/frai.2023.1169595.
378
379 29. Day T. A Preliminary Investigation of Fake Peer-Reviewed Citations and References Generated by ChatGPT. The Professional
380 Geographer. 2023 Mar 23:1-4. DOI: 10.1080/00330124.2023.2190373.
381
382 30. de Oliveira RS, Ballestero M. The future of Pediatric Neurosurgery and ChatGPT: an editor's perspective. Archives of Pediatric
383 Neurosurgery 2023;5. DOI: https://doi.org/10.46900/apn.v5i2.191.
384
385 31. De Vito EL. Artificial intelligence and chatGPT. Would you read an artificial author? Medicina 2023;83:329-332.
386
387 32. Dergaa I, Chamari K, Zmijewski P, Ben Saad H. From human writing to artificial intelligence generated text: examining the prospects and
388 potential threats of ChatGPT in academic writing. Biol Sport 2023;40:615-622.
389
78
390 33. Donato H, Escada P, Villanueva T. The Transparency of Science with ChatGPT and the Emerging Artificial Intelligence Language Models:
391 Where Should Medical Journals Stand? Acta Med Port. 2023 Mar 1;36(3):147-148. doi: 10.20344/amp.19694. Epub 2023 Feb 9.
392
393 34. Dunn C, Hunter J, Steffes W, Whitney Z, Foss M, Mammino J, Leavitt A, Hawkins SD, Dane A, Yungmann M, Nathoo R. Artificial
394 intelligence-derived dermatology case reports are indistinguishable from those written by humans: A single-blinded observer study. J Am
395 Acad Dermatol. 2023 Apr 11:S0190-9622(23)00587-X. doi: 10.1016/j.jaad.2023.04.005. Epub ahead of print.
396
397 35. Fatani B. ChatGPT for Future Medical and Dental Research. Cureus. 2023 Apr 8;15(4):e37285. doi: 10.7759/cureus.37285.
398
399 36. Galland J. Les chatbots en médecine interne : opportunités et défis à venir [Chatbots and internal medicine: Future opportunities and
400 challenges]. Rev Med Interne. 2023 May;44(5):209-211. French. doi: 10.1016/j.revmed.2023.04.001. Epub 2023 Apr 29.
401
402 37. Gandhi Periaysamy A, Satapathy P, Neyazi A, Padhi BK. ChatGPT: roles and boundaries of the new artificial intelligence tool in medical
403 education and health research - correspondence. Ann Med Surg (Lond) 2023;85:1317-1318.
404
405 38. Gao CA, Howard FM, Markov NS, Dyer EC, Ramesh S, Luo Y, Pearson AT. Comparing scientific abstracts generated by ChatGPT to
406 original abstracts using an artificial intelligence output detector, plagiarism detector, and blinded human reviewers. bioRxiv
407 2022.12.23.521610; doi: https://doi.org/10.1101/2022.12.23.521610.
408
409 39. Goedde D, Noehl S, Wolf C, Rupert Y, Rimkus L, Ehlers J, Breuckmann F, Sellmann T. ChatGPT in medical literature-a concise review and
410 SWOT analysis. medRxiv 2023.05.06.23289608; doi: https://doi.org/10.1101/2023.05.06.23289608.
411
412 40. Gordijn B, Have HT. ChatGPT: evolution or revolution? Med Health Care Philos. 2023 Mar;26(1):1-2. doi: 10.1007/s11019-023-10136-0.
413
414 41. Gottlieb M, Kline JA, Schneider AJ, Coates WC. ChatGPT and conversational artificial intelligence: Friend, foe, or future of research? Am J
415 Emerg Med. 2023 May 18;70:81-83. doi: 10.1016/j.ajem.2023.05.018. Epub ahead of print.
416
417 42. Graf A, Bernardi RE. ChatGPT in Research: Balancing Ethics, Transparency and Advancement. Neuroscience. 2023 Apr 1;515:71-73. doi:
418 10.1016/j.neuroscience.2023.02.008. Epub 2023 Feb 21.
419
420 43. Graham A. ChatGPT and other AI tools put students at risk of plagiarism allegations, MDU warns. BMJ. 2023 May 17;381:1133. doi:
421 10.1136/bmj.p1133.
422
423 44. Gravel J, D'Amours-Gravel M, Osmanlliu E. Learning to fake it: limited responses and fabricated references provided by ChatGPT for
424 medical questions. medRxiv 2023.03.16.23286914; doi: https://doi.org/10.1101/2023.03.16.23286914.
79
425
426 45. Guo E, Gupta M, Sinha S, Rössler K, Tatagiba M, Akagami R, Al-Mefty O, Sugiyama T, Stieg PE, Pickett GE, de Lotbiniere-Bassett M.
427 neuroGPT-X: Towards an Accountable Expert Opinion Tool for Vestibular Schwannoma. medRxiv 2023.02.25.23286117; doi:
428 https://doi.org/10.1101/2023.02.25.23286117.
429
430 46. Gurha P, Ishaq N, Marian AJ. ChatGPT and other artificial intelligence chatbots and biomedical writing. J Cardiovasc Aging. 2023;3(2):20.
431 doi: 10.20517/jca.2023.13. Epub 2023 Mar 31.
432
433 47. Haemmerli J, Sveikata L, Nouri A, May A, Egervari K, Freyschlag C, Lobrinus JA, Migliorini D, Momjian S, Sanda N, Schaller K. ChatGPT in
434 glioma patient adjuvant therapy decision making: ready to assume the role of a doctor in the tumour board?. medRxiv
435 2023.03.19.23287452; doi: https://doi.org/10.1101/2023.03.19.23287452.
436
437 48. Harskamp RE, De Clercq L. Performance of ChatGPT as an AI-assisted decision support tool in medicine: a proof-of-concept study for
438 interpreting symptoms and management of common cardiac conditions (AMSTELHEART-2). medRxiv 2023.03.25.23285475; doi:
439 https://doi.org/10.1101/2023.03.25.23285475.
440
441 49. Hill-Yardin EL, Hutchinson MR, Laycock R, Spencer SJ. A Chat(GPT) about the future of scientific publishing. Brain Behav Immun. 2023
442 May;110:152-154. doi: 10.1016/j.bbi.2023.02.022. Epub 2023 Mar 1.
443
444 50. Hirani R, Farabi B, Marmon S. Experimenting with ChatGPT: Concerns for academic medicine. J Am Acad Dermatol. 2023 May 11:S0190-
445 9622(23)00747-8. doi: 10.1016/j.jaad.2023.04.045. Epub ahead of print.
446
447 51. Homolak J. Opportunities and risks of ChatGPT in medicine, science, and academic publishing: a modern Promethean dilemma. Croat Med
448 J. 2023 Feb 28;64(1):1-3. doi: 10.3325/cmj.2023.64.1.
449
450 52. Hosseini M, Gao CA, Liebovitz DM, Carvalho AM, Ahmad FS, Luo Y, MacDonald N, Holmes KL, Kho A. An exploratory survey about using
451 ChatGPT in education, healthcare, and research. medRxiv 2023.03.31.23287979; doi: https://doi.org/10.1101/2023.03.31.23287979.
452
453 53. Howard A, Hope W, Gerada A. ChatGPT and antimicrobial advice: the end of the consulting infection doctor? Lancet Infect Dis. 2023
454 Apr;23(4):405-406. doi: 10.1016/S1473-3099(23)00113-5. Epub 2023 Feb 20.
455
456 54. Hsu TW, Tsai SJ, Ko CH, Thompson T, Hsu CW, Yang FC, Tsai CK, Tu YK, Yang SN, Tseng PT, Liang CS. Plagiarism, Quality, and
457 Correctness of ChatGPT-Generated vs Human-Written Abstract for Research Paper. Available at SSRN:
458 https://ssrn.com/abstract=4429014 or http://dx.doi.org/10.2139/ssrn.4429014
459
80
460 55. Huang J, Tan M. The role of ChatGPT in scientific communication: writing better scientific review articles. Am J Cancer Res. 2023 Apr
461 15;13(4):1148-1154.
462
463 56. Hurley D. Your AI Program Will Write Your Paper Now: Neurology Editors on Managing Artificial Intelligence Submissions. Neurology Today
464 2023;23:10-11.
465
466 57. Janssen BV, Kazemier G, Besselink MG. The use of ChatGPT and other large language models in surgical science. BJS Open. 2023 Mar
467 7;7(2):zrad032. doi: 10.1093/bjsopen/zrad032.
468
469 58. Johnson SB, King AJ, Warner EL, Aneja S, Kann BH, Bylund CL. Using ChatGPT to evaluate cancer myths and misconceptions: artificial
470 intelligence and cancer information. JNCI Cancer Spectr. 2023 Mar 1;7(2):pkad015. doi: 10.1093/jncics/pkad015.
471
472 59. Juhi A, Pipil N, Santra S, Mondal S, Behera JK, Mondal H. The Capability of ChatGPT in Predicting and Explaining Common Drug-Drug
473 Interactions. Cureus. 2023 Mar 17;15(3):e36272. doi: 10.7759/cureus.36272.
474
475 60. Kaneda Y. In the Era of Prominent AI, What Role Will Physicians Be Expected to Play? QJM. 2023 May 22:hcad099. doi:
476 10.1093/qjmed/hcad099. Epub ahead of print.
477
478 61. Kim J. Search for Medical Information and Treatment Options for Musculoskeletal Disorders through an Artificial Intelligence Chatbot:
479 Focusing on Shoulder Impingement Syndrome. medRxiv 2022.12.16.22283512; doi: https://doi.org/10.1101/2022.12.16.22283512.
480
481 62. Kim SG. Using ChatGPT for language editing in scientific articles. Maxillofac Plast Reconstr Surg. 2023 Mar 8;45(1):13. doi:
482 10.1186/s40902-023-00381-x.
483
484 63. Koo M. The Importance of Proper Use of ChatGPT in Medical Writing. Radiology. 2023 May;307(3):e230312. doi: 10.1148/radiol.230312.
485 Epub 2023 Mar 7.
486
487 64. Kumar AH. Analysis of ChatGPT Tool to Assess the Potential of its Utility for Academic Writing in Biomedical Domain. BEMS Reports
488 [Internet]. 2023Feb.2 [cited 2023Jun.10];9(1):24-30. Available from: https://bemsreports.org/index.php/bems/article/view/132.
489
490 65. Lee P, Bubeck S, Petro J. Benefits, Limits, and Risks of GPT-4 as an AI Chatbot for Medicine. N Engl J Med. 2023 Mar 30;388(13):1233-
491 1239. doi: 10.1056/NEJMsr2214184.
492
493 66. Levin G, Meyer R, Yasmeen A, Young B, Guige PA, Bar-Noy T, Tatar A, Perelstein O, Brezinov Y. ChatGPT-written OBGYN abstracts fool
494 practitioners. Am J Obstet Gynecol MFM. 2023 Apr 29:100993. doi: 10.1016/j.ajogmf.2023.100993. Epub ahead of print.
81
495
496 67. Li H, Moon JT, Purkayastha S, Celi LA, Trivedi H, Gichoya JW. Ethics of large language models in medicine and medical research. Lancet
497 Digit Health. 2023 Jun;5(6):e333-e335. doi: 10.1016/S2589-7500(23)00083-3. Epub 2023 Apr 27.
498
499 68. Li S. ChatGPT has made the field of surgery full of opportunities and challenges. Int J Surg. 2023 May 17. doi:
500 10.1097/JS9.0000000000000454. Epub ahead of print.
501
502 69. Liebrenz M, Schleifer R, Buadze A, Bhugra D, Smith A. Generating scholarly content with ChatGPT: ethical challenges for medical
503 publishing. Lancet Digit Health. 2023 Mar;5(3):e105-e106. doi: 10.1016/S2589-7500(23)00019-5. Epub 2023 Feb 6.
504
505 70. Lin Z. Modernizing authorship criteria: Challenges from exponential authorship inflation and generative artificial intelligence. 2023.
506 (Preprint). https://psyarxiv. com/s6h58.
507
508 71. Loh E. ChatGPT and generative AI chatbots: challenges and opportunities for science, medicine and medical leaders. BMJ Lead. 2023 May
509 2:leader-2023-000797. doi: 10.1136/leader-2023-000797. Epub ahead of print.
510
511 72. Maeker E, Maeker-Poquet B. ChatGPT : une solution pour rédiger des revues de littérature en médecine ?, Volume 8035, Issue 135,
512 06/2023, Pages 137-212, ISSN 1627-4830, http://dx.doi.org/10.1016/j.npg.2023.03.002
513 (http://www.sciencedirect.com/science/article/pii/S1627-4830(23)00037-5).
514
515 73. Marchandot B, Matsushita K, Carmona A, Trimaille A, Morel O. ChatGPT: the next frontier in academic writing for cardiologists or a
516 pandora's box of ethical dilemmas. Eur Heart J Open. 2023 Feb 13;3(2):oead007. doi: 10.1093/ehjopen/oead007.
517
518 74. Martínez-Sellés M, Marina-Breysse M. Current and Future Use of Artificial Intelligence in Electrocardiography. J Cardiovasc Dev Dis. 2023
519 Apr 17;10(4):175. doi: 10.3390/jcdd10040175.
520
521 75. Mehnen L, Gruarin S, Vasileva M, Knapp B. ChatGPT as a medical doctor? A diagnostic accuracy study on common and rare diseases.
522 medRxiv 2023.04.20.23288859; doi: https://doi.org/10.1101/2023.04.20.23288859.
523
524 76. Mello MM, Guha N. ChatGPT and Physicians' Malpractice Risk. JAMA Health Forum. 2023 May 5;4(5):e231938. doi:
525 10.1001/jamahealthforum.2023.1938.
526
527 77. Mese I. The imperative of a radiology AI deployment registry and the potential of ChatGPT. Clin Radiol. 2023 Jul;78(7):554. doi:
528 10.1016/j.crad.2023.04.001. Epub 2023 Apr 18.
529
82
530 78. Nastasi NJ, Courtright KR, Halpern SD, Weissman GE. Does ChatGPT Provide Appropriate and Equitable Medical Advice? A Vignette-
531 Based, Clinical Evaluation Across Care Contexts. medRxiv 2023.02.25.23286451; doi: https://doi.org/10.1101/2023.02.25.23286451.
532
533 79. Nguyen Y, Costedoat-Chalumeau N. Les intelligences artificielles conversationnelles en médecine interne : l’exemple de
534 l’hydroxychloroquine selon ChatGPT [Artificial intelligence and internal medicine: The example of hydroxychloroquine according to
535 ChatGPT]. Rev Med Interne. 2023 May;44(5):218-226. French. doi: 10.1016/j.revmed.2023.03.017.
536
537 80. Nógrádi B, Polgár TF, Meszlényi V, Kádár Z, Hertelendy P, Csáti A, et al. Is There Any Room for Generative AI in Neurology and Other
538 Medical Areas?. Available at SSRN: https://ssrn.com/abstract=4372965 or http://dx.doi.org/10.2139/ssrn.4372965.
539
540 81. North RA. Plagiarism Reimagined. Function. 2023; 4(3):zqad014, https://doi.org/10.1093/function/zqad014.
541
542 82. Oh N, Choi GS, Lee WY. ChatGPT goes to the operating room: evaluating GPT-4 performance and its potential in surgical education and
543 training in the era of large language models. Ann Surg Treat Res. 2023 May;104(5):269-273. doi: 10.4174/astr.2023.104.5.269. Epub 2023
544 Apr 28.
545
546 83. Okan Ç. AI and Psychiatry: The ChatGPT Perspective. Alpha Psychiatry. 2023 Mar 1;24(2):41-42. doi:
547 10.5152/alphapsychiatry.2023.010223.
548
549 84. Parsa A, Ebrahimzadeh MH. ChatGPT in Medicine; a Disruptive Innovation or Just One Step Forward? Arch Bone Jt Surg. 2023;11(4):225-
550 226. doi: 10.22038/abjs.2023.22042.
551
552 85. Patel SB, Lam K, Liebrenz M. ChatGPT: friend or foe. Lancet Digit Health 2023;5:e102. DOI: https://doi.org/10.1016/S2589-7500(23)00023-
553 7.
554
555 86. Pourhoseingholi MA, Hatamnejad MR, Solhpour A. Does chatGPT (or any other artificial intelligence language tool) deserve to be included
556 in authorship list? Gastroenterol Hepatol Bed Bench. 2023;16(1):435-437. doi: 10.22037/ghfbb.v16i1.2747.
557
558 87. Rao D. The Urgent Need for Healthcare Workforce Upskilling and Ethical Considerations in the Era of AI-Assisted Medicine. Indian J
559 Otolaryngol Head Neck Surg (2023). https://doi.org/10.1007/s12070-023-03755-9.
560
561 88. Ray PP, Majumder P. AI Tackles Pandemics: ChatGPT's Game-Changing Impact on Infectious Disease Control. Ann Biomed Eng. 2023
562 May 18:1–3. doi: 10.1007/s10439-023-03239-5. Epub ahead of print.
563
83
564 89. Ros-Arlanzón P, Pérez-Sempere A. ChatGPT: una novedosa herramienta de escritura para artículos científicos, pero no un autor (por el
565 momento) [ChatGPT: a novel tool for writing scientific articles, but not an author (for the time being)]. Rev Neurol. 2023 Apr 16;76(8):277.
566 Spanish. doi: 10.33588/rn.7608.2023066.
567
568 90. Sabry Abdel-Messih M, Kamel Boulos MN. ChatGPT in Clinical Toxicology. JMIR Med Educ. 2023 Mar 8;9:e46876. doi: 10.2196/46876.
569
570 91. Sallam M. The Utility of ChatGPT as an Example of Large Language Models in Healthcare Education, Research and Practice: Systematic
571 Review on the Future Perspectives and Potential Limitations. medRxiv 2023.02.19.23286155; doi:
572 https://doi.org/10.1101/2023.02.19.23286155.
573
574 92. Sallam M, Salim NA, Al-Tammemi AB, Barakat M, Fayyad D, Hallit S, Harapan H, Hallit R, Mahafzah A. ChatGPT Output Regarding
575 Compulsory Vaccination and COVID-19 Vaccine Conspiracy: A Descriptive Study at the Outset of a Paradigm Shift in Online Search for
576 Information. Cureus. 2023 Feb 15;15(2):e35029. doi: 10.7759/cureus.35029.
577
578 93. Salvagno M, Taccone FS, Gerli AG. Can artificial intelligence help for scientific writing? Crit Care. 2023 Feb 25;27(1):75. doi:
579 10.1186/s13054-023-04380-2. Erratum in: Crit Care. 2023 Mar 8;27(1):99.
580
581 94. Sanmarchi F, Bucci A, Nuzzolese AG. et al. A step-by-step researcher's guide to the use of an AI-based transformer in epidemiology: an
582 exploratory analysis of ChatGPT using the STROBE checklist for observational studies. J Public Health (Berl.) (2023).
583 https://doi.org/10.1007/s10389-023-01936-y
584
585 95. Sarink MJ, Bakker IL, Anas AA, Yusuf E. A study on the performance of ChatGPT in infectious diseases clinical consultation. Clin Microbiol
586 Infect. 2023 May 18:S1198-743X(23)00241-0. doi: 10.1016/j.cmi.2023.05.017. Epub ahead of print.
587
588 96. Schulte B. Capacity of ChatGPT to Identify Guideline-Based Treatments for Advanced Solid Tumors. Cureus. 2023 Apr 21;15(4):e37938.
589 doi: 10.7759/cureus.37938.
590
591 97. Singh OP. Artificial intelligence in the era of ChatGPT - Opportunities and challenges in mental health care. Indian J Psychiatry. 2023
592 Mar;65(3):297-298. doi: 10.4103/indianjpsychiatry.indianjpsychiatry_112_23.
593
594 98. Singh S, Djalilian A, Ali MJ. ChatGPT and Ophthalmology: Exploring Its Potential with Discharge Summaries and Operative Notes. Semin
595 Ophthalmol. 2023 May 3:1-5. doi: 10.1080/08820538.2023.2209166. Epub ahead of print.
596
597 99. Tang L, Sun Z, Idnay B, Nestor JG, Soroush A, Elias PA, Xu Z, Ding Y, Durrett G, Rousseau J, Weng C, Peng Y. Evaluating Large
598 Language Models on Medical Evidence Summarization. medRxiv 2023.04.22.23288967; doi: https://doi.org/10.1101/2023.04.22.23288967.
84
599
600 100. Temsah O, Khan SA, Chaiah Y, Senjab A, Alhasan K, Jamal A, Aljamaan F, Malki KH, Halwani R, Al-Tawfiq JA, Temsah MH, Al-
601 Eyadhy A. Overview of Early ChatGPT's Presence in Medical Literature: Insights From a Hybrid Literature Review by ChatGPT and Human
602 Experts. Cureus. 2023 Apr 8;15(4):e37281. doi: 10.7759/cureus.37281.
603
604 101. Thorp HH. ChatGPT is fun, but not an author. Science. 2023 Jan 27;379(6630):313. doi: 10.1126/science.adg7879. Epub 2023 Jan 26.
605
606 102. Haq ZU, Naeem H, Naeem A, Iqbal F, Zaeem D. Comparing human and artificial intelligence in writing for health journals: an
607 exploratory study. medRxiv 2023.02.22.23286322; doi: https://doi.org/10.1101/2023.02.22.23286322.
608
609 103. Uprety D, Zhu D, West HJ. ChatGPT-A promising generative AI tool and its implications for cancer care. Cancer. 2023 May 14. doi:
610 10.1002/cncr.34827. Epub ahead of print.
611
612 104. Uz C, Umay E. "Dr ChatGPT": Is it a reliable and useful source for common rheumatic diseases? Int J Rheum Dis. 2023 May 23. doi:
613 10.1111/1756-185X.14749. Epub ahead of print.
614
615 105. van Dis EAM, Bollen J, Zuidema W, van Rooij R, Bockting CL. ChatGPT: five priorities for research. Nature. 2023 Feb;614(7947):224-
616 226. doi: 10.1038/d41586-023-00288-7.
617
618 106. Waisberg E, Ong J, Masalkhi M, Kamran SA, Zaman N, Sarker P, Lee AG, Tavakkoli A. GPT-4: a new era of artificial intelligence in
619 medicine. Ir J Med Sci. 2023 Apr 19. doi: 10.1007/s11845-023-03377-8. Epub ahead of print.
620
621 107. Wen J, Wang W. The future of ChatGPT in academic research and publishing: A commentary for clinical and translational medicine.
622 Clin Transl Med. 2023 Mar;13(3):e1207. doi: 10.1002/ctm2.1207.
623
624 108. Xue VW, Lei P, Cho WC. The potential impact of ChatGPT in clinical and translational medicine. Clin Transl Med. 2023
625 Mar;13(3):e1216. doi: 10.1002/ctm2.1216.
626
627 109. Yadava OP. ChatGPT—a foe or an ally?. Indian J Thorac Cardiovasc Surg 2023;39:217–21. https://doi.org/10.1007/s12055-023-01507-
628 6.
629
630 110. Au Yeung J, Kraljevic Z, Luintel A, Balston A, Idowu E, Dobson RJ, Teo JT. AI chatbots not yet ready for clinical use. Front Digit Health.
631 2023 Apr 12;5:1161098. doi: 10.3389/fdgth.2023.1161098.
632
85
633 111. Young JN, O'Hagan R, Poplausky D, et al. The utility of ChatGPT in generating patient-facing and clinical responses for melanoma. J
634 Am Acad Dermatol. 2023 May 18:S0190-9622(23)00908-8. doi: 10.1016/j.jaad.2023.05.024.
635
636 112. Zheng H, Zhan H. ChatGPT in Scientific Writing: A Cautionary Tale. Am J Med. 2023 Mar 10:S0002-9343(23)00159-6. doi:
637 10.1016/j.amjmed.2023.02.011. Epub ahead of print.
638
639 113. Zhong Y, Chen YJ, Zhou Y, Lyu YA, Yin JJ, Gao YJ. The Artificial intelligence large language models and neuropsychiatry practice and
640 research ethic. Asian J Psychiatr 2023;84:103577.
641
642 114. Zhou J, Jia Y, Qiu Y, Lin L. The Potential of Applying Chatgpt to Extract Keywords of Medical Literature in Plastic Surgery. Aesthet Surg
643 J. 2023 May 20:sjad158. doi: 10.1093/asj/sjad158. Epub ahead of print.
644
645 115. Zhou Z. Evaluation of ChatGPT's Capabilities in Medical Report Generation. Cureus 2023;15:e37589.
646
647 116. Zhu L, Mou W, Chen R. Can the ChatGPT and other large language models with internet-connected database solve the questions and
648 concerns of patient with prostate cancer and help democratize medical knowledge? J Transl Med. 2023 Apr 19;21(1):269.
649
650 117. Zielinski C, Winker M, Aggarwal R, et al. Chatbots, ChatGPT, and Scholarly Manuscripts: WAME Recommendations on ChatGPT and
651 Chatbots in Relation to Scholarly Publications. Open Access Macedonian Journal of Medical Sciences 2023;11:83-86. DOI:
652 10.3889/oamjms.2023.11502.
653
654 118. Zimmerman A. A Ghostwriter for the Masses: ChatGPT and the Future of Writing. Ann Surg Oncol. 2023 Jun;30(6):3170-3173.
655
656
657
658
659
660
661
662
663
664
665
666
667
86
668
669
670
671
672
673
674 Table-2: Summaries of full text of 118 selected article were uploaded in pdf format on ChatPDF, subsequently 12
675 questions were asked. Questions and respective answers given by ChatPDF are listed in this table.
676
677
Number Question Answer by ChatPDF Summary
1 What are potential ChatGPT can be used in medical writing to produce coherent and ChatGPT has
role of ChatGPT in readable content. It can also be used in scientific research to generate applications in
medical writing creative ideas, write essays and homework assignments, and even medical writing,
and research? contribute as an author on research papers. Additionally, ChatGPT has scientific
the potential to automate medical tasks and enhance learning, research, and
interpretation, and recall of medical information. However, concerns automation of
persist about its accuracy, potential biases in the training data, privacy medical tasks, but
concerns, issues related to authorship in academic papers, and ethical concerns remain
considerations of its use. Despite these challenges, AI tools like regarding
ChatGPT are expected to be widely adopted in the medical field in the accuracy, biases,
future. privacy,
authorship, and
ethics. Adoption
in the medical
field is
anticipated.
2 What could be the ChatGPT can assist with patient inquiries, help with writing medical notes ChatGPT offers
role of ChatGPT in and discharge summaries, aid in decision-making and treatment assistance in
clinical practice? planning, and support medical educators in various tasks. In clinical patient inquiries,
practice, ChatGPT could potentially be used as a valuable tool for medical note
personalized learning and encouraging a shift towards critical thinking writing, decision-
87
and problem-based learning. However, it is important to note that there making, and
are challenges associated with the use of ChatGPT in clinical practice, medical
such as the need for access to current information, potential biases in the education, but
training data, privacy concerns, issues related to authorship in academic challenges like
papers, and ethical considerations of its use. Therefore, regulations and biases, privacy,
control mechanisms should be established to address ethical utilization of authorship, and
ChatGPT and similar tools in medicine. ethics need to be
addressed for
ethical utilization
in medicine.
3 What are ethical Paper writing entails several ethical concerns, including the risks of bias, Ethical concerns
issues associated plagiarism, copyright infringement, transparency issues, legal in paper writing
with paper implications, lack of originality, incorrect responses, limited knowledge, with ChatGPT
writing? and inaccurate citations. These concerns arise from the potential for involve bias,
ChatGPT to generate fabricated and unreliable research outputs, as well plagiarism,
as the lack of transparency and accountability in its responses. transparency,
Addressing these ethical issues is crucial to uphold research integrity and legal issues, and
ensure the reliability and trustworthiness of the generated results. lack of
accountability,
which need
addressing for
reliable and
trustworthy
research.
4 Can CHATGPT be The question of whether ChatGPT can be considered an author is a Including
an author? subject of controversy and may pose challenges in a legal context. The ChatGPT as an
traditional authorship criteria, such as those outlined by the International author is
Committee of Medical Journal Editors (ICMJE), require individuals to controversial due
contribute substantially to various aspects of the research process, to its inability to
including conception, design, analysis, interpretation, and manuscript meet authorship
preparation. As an AI language model, ChatGPT cannot independently criteria, but it can
fulfill these criteria. be used as a tool
88
to assist human
However, ChatGPT can serve as a valuable tool to support human authors. Decisions
authors in writing papers and generating ideas. Its contributions can be should be made
acknowledged in the paper, but it may not meet the requirements for on a case-by-case
authorship in the traditional sense. basis, adhering to
ethical guidelines.
Determining whether or not to include ChatGPT as an author should be
approached on a case-by-case basis and aligned with established ethical
guidelines for scientific research. It is essential to consider the specific
contributions made by ChatGPT and ensure that human authors take
responsibility for the substantial aspects of the research process as
outlined by the ICMJE guidelines.
5 Can CHATGPT ChatGPT can assist in the editing process, including language editing, ChatGPT can aid
write text in good and summarizing the text to compose a suitable abstract. However, the in the editing
English and free quality of the AI-generated content needs careful evaluation as it is not a process, but
of plagiarism? substitute for human expertise, creativity, and critical thinking. While human expertise,
ChatGPT can rapidly comprehend information and connect evidence, it creativity, and
lacks the nuance, style, or originality that a human author can provide. critical thinking
Additionally, there is a risk of plagiarism if ChatGPT is not properly are essential.
trained or if it draws on sources without proper attribution. Therefore, it is Careful evaluation
important to carefully evaluate and edit any content generated by and editing are
ChatGPT to ensure that it meets high standards of quality and originality. needed to ensure
quality, originality,
and avoid
plagiarism.
6 Role of CHATGPT While there is limited specific information on the role of ChatGPT in While there is no
so far in neurological disorders related to clinical practice and research, it is worth specific
neurological noting that the integration of artificial intelligence, natural language information on
disorders related processing, and GPT technologies in psychiatry shows promise in ChatGPT's role in
clinical practice transforming the diagnosis, treatment, and monitoring of psychiatric neurological
and research. disorders. These technologies have the potential to assist psychiatrists in disorders, AI
making more accurate diagnoses, developing personalized treatment technologies
89
plans, enhancing the accuracy and efficiency of clinical trials, and show promise in
increasing access to mental health care through the use of chatbots and revolutionizing
virtual assistants. Although there may not be explicit examples or studies psychiatric
on ChatGPT's application in neurological disorders, similar applications diagnosis,
could potentially be developed in the future. treatment, and
accessibility to
mental health
care.
7 Effectiveness and Upon review, I couldn't find specific information on the effectiveness and There is no
efficiency of Chat efficiency of ChatGPT in medical research and clinical settings. The specific
GPT in medical provided pages discuss the potential applications of AI and natural information on
research and language processing in medicine, including psychiatry, but do not offer ChatGPT's
clinical settings specific studies or data on the use of ChatGPT in these settings. effectiveness in
medical research
It is important to emphasize that while AI tools like ChatGPT show and clinical
promise in improving medical research and clinical practice, they should settings,
not be seen as a substitute for human expertise. Instead, they should be highlighting the
utilized alongside established best practices and ethical guidelines. need for further
Further research is necessary to assess the effectiveness and efficiency research and
of ChatGPT specifically in medical research and clinical contexts. consideration of
ethical guidelines.
8 Potential benefits ChatGPT presents potential benefits and limitations in medical research ChatGPT has
and limitations of and clinical applications. Some benefits include its assistance in patient potential benefits
Chat GPT in questionnaires, interviews, and epidemiological research by streamlining in mental health
medical research data collection, recording, and analysis. The integration of artificial care, data
and clinical intelligence, natural language processing, and GPT technologies in collection, and
applications psychiatry holds promise for transforming psychiatric diagnosis, manuscript
treatment, and monitoring. It can aid in accurate diagnoses, personalized checks but has
treatment design, improved clinical trial efficiency, and increased limitations in
accessibility to mental health care through chatbots and virtual providing
assistants. ChatGPT can also be useful for tasks like proofreading and comprehensive
manuscript checks. diagnoses and
90
replacing human
However, there are limitations to consider. ChatGPT cannot update its expertise,
training data in real-time and may provide general or vague answers in creativity, and
medical-related conversations. Its diagnoses may not be comprehensive critical thinking.
or sufficient, indicating that it cannot independently handle the Careful evaluation
complexities of clinical practice. It cannot replace the specialized and ethical
knowledge, innovative ideas, and opinions that human input provides. guidelines are
Ethical guidelines and further exploration of research ethics are necessary.
necessary for the responsible use of AI in health research. The quality of
AI-generated content requires careful evaluation as it cannot substitute
for human expertise, creativity, and critical thinking. While ChatGPT can
comprehend information and connect evidence rapidly, it lacks the
nuance and originality that human authors provide. Moreover, there is a
risk of plagiarism if ChatGPT is not appropriately trained or if it fails to
provide proper source attribution.
9 The ethical There are ethical implications of using ChatGPT in medical research and Using ChatGPT in
implications of clinical practice. - Almost 97% of the records expressed concerns or medical research
using Chat GPT in possible risks associated with ChatGPT. Ethical issues, such as risk of and clinical
medical research bias, plagiarism, copyright issues, transparency issues, legal concerns, practice raises
and clinical lack of originality, incorrect responses, limited knowledge, and inaccurate ethical concerns
practice citations were the most common. The quality of the AI-generated content such as bias,
needs careful evaluation as it is not a substitute for human expertise, plagiarism, and
creativity, and critical thinking. While ChatGPT can rapidly comprehend limited human-like
information and connect evidence, it lacks the nuance or originality that a qualities.
human author can provide. Additionally, there is a risk of plagiarism if Adherence to
ChatGPT is not properly trained or if it draws on sources without proper research ethics
attribution. Further exploration of research ethics and the moral use of AI and careful
in health research is required to establish guidelines. It is important to evaluation are
consider these ethical implications when using ChatGPT in medical crucial.
research and clinical practice. Researchers and practitioners should
ensure that they are using AI tools like ChatGPT in an ethical manner
that respects patient privacy and autonomy while also adhering to
91
medical research datasets. 2. Clinical trial recruitment: ChatGPT could help identify eligible and clinical
and clinical patients for clinical trials by analyzing electronic health records and other practice include
practice data sources. 3. Clinical data management: ChatGPT could help manage clinical decision
large amounts of clinical data by extracting relevant information from support, trial
electronic health records and other sources. 4. Research support: recruitment, data
ChatGPT could assist researchers in identifying relevant literature, management,
generating hypotheses, and analyzing data. 5. Patient education: research support,
ChatGPT could provide personalized education to patients about their and patient
conditions, treatments, and medications. 6. Automation of tasks: education. Ethical
ChatGPT could automate tasks such as data analysis and image considerations
interpretation, freeing up time for healthcare professionals to focus on and effective
patient care. 7. Drug discovery: ChatGPT could assist in drug discovery utilization are
by recognizing, classifying, and describing chemical formulas or crucial.
molecular structures. 8. Disease prediction, diagnosis, and assessment
of therapeutic targets: ChatGPT could help predict diseases based on
patient data, diagnose diseases based on symptoms or test results, and
assess therapeutic targets for drug development. Overall, the potential
future applications of ChatGPT in medical research and clinical practice
are vast. However, it is important to address ethical concerns associated
with its use while also ensuring that it is used effectively to improve
patient outcomes.
12 Recommendations 1. Researchers should be transparent about the use of ChatGPT in Researchers,
for researchers, scientific publications and ensure that the generated text is accurate and clinicians, and
clinicians, and reliable. policymakers
policymakers on 2. Clinicians should use ChatGPT with caution and recognize its should ensure
the use of limitations in providing reliable information. It can be useful for tasks like transparency,
ChatGPT in proofreading and manuscript checks, but cannot replace the specialized evaluate
medical research knowledge, innovative ideas, and opinions that human input provides. limitations,
and clinical 3. Policymakers should establish regulations and control mechanisms to establish
practice address ethical utilization of ChatGPT and similar tools in medicine. This regulations, and
includes guidelines for transparency in disclosing the use of AI tools like address ethical
ChatGPT in scientific publications. concerns related
93