101
|
Pal S, Bhattacharya M, Islam MA, Chakraborty C. AI-enabled ChatGPT or LLM: a new algorithm is required for plagiarism-free scientific writing. Int J Surg 2024; 110:1329-1330. [PMID: 38000076 PMCID: PMC10871629 DOI: 10.1097/js9.0000000000000939] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2023] [Accepted: 11/09/2023] [Indexed: 11/26/2023]
Affiliation(s)
- Soumen Pal
- School of Mechanical Engineering, Vellore Institute of Technology, Vellore, Tamil Nadu, India
| | - Manojit Bhattacharya
- Department of Zoology, Fakir Mohan University, Vyasa Vihar, Balasore, Odisha, India
| | - Md. Aminul Islam
- Advanced Molecular Lab, Department of Microbiology, President Abdul Hamid Medical College, Karimganj, Kishoreganj
- COVID-19 Diagnostic Lab, Department of Microbiology, Noakhali Science and Technology University, Noakhali, Bangladesh
| | - Chiranjib Chakraborty
- Department of Biotechnology, School of Life Science and Biotechnology, Adamas University, Kolkata, West Bengal, India
| |
Collapse
|
102
|
Vieira AGDS, Saconato H, Eid RAC, Nawa RK. ChatGPT: immutable insertion in health research and researchers' lives. EINSTEIN-SAO PAULO 2024; 22:eCE0752. [PMID: 38477797 PMCID: PMC11730319 DOI: 10.31744/einstein_journal/2024ce0752] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Accepted: 10/18/2023] [Indexed: 03/14/2024] Open
|
103
|
Shah PS, Acharya G. Artificial intelligence/machine learning and journalology: Challenges and opportunities. Acta Obstet Gynecol Scand 2024; 103:196-198. [PMID: 38284152 PMCID: PMC10823383 DOI: 10.1111/aogs.14772] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2023] [Accepted: 12/16/2023] [Indexed: 01/30/2024]
Affiliation(s)
- Prakesh S. Shah
- Department of PediatricsMount Sinai Hospital, University of TorontoTorontoOntarioCanada
| | - Ganesh Acharya
- Department of Clinical Science, Intervention and Technology (CLINTEC)Karolinska Institutet and Center for Fetal Medicine, Karolinska University HospitalStockholmSweden
- Women's Health and Perinatology Research GroupUiT – The Arctic University of NorwayTromsøNorway
| |
Collapse
|
104
|
Kacena MA, Plotkin LI, Fehrenbacher JC. The Use of Artificial Intelligence in Writing Scientific Review Articles. Curr Osteoporos Rep 2024; 22:115-121. [PMID: 38227177 PMCID: PMC10912250 DOI: 10.1007/s11914-023-00852-0] [Citation(s) in RCA: 14] [Impact Index Per Article: 14.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 12/21/2023] [Indexed: 01/17/2024]
Abstract
PURPOSE OF REVIEW With the recent explosion in the use of artificial intelligence (AI) and specifically ChatGPT, we sought to determine whether ChatGPT could be used to assist in writing credible, peer-reviewed, scientific review articles. We also sought to assess, in a scientific study, the advantages and limitations of using ChatGPT for this purpose. To accomplish this, 3 topics of importance in musculoskeletal research were selected: (1) the intersection of Alzheimer's disease and bone; (2) the neural regulation of fracture healing; and (3) COVID-19 and musculoskeletal health. For each of these topics, 3 approaches to write manuscript drafts were undertaken: (1) human only; (2) ChatGPT only (AI-only); and (3) combination approach of #1 and #2 (AI-assisted). Articles were extensively fact checked and edited to ensure scientific quality, resulting in final manuscripts that were significantly different from the original drafts. Numerous parameters were measured throughout the process to quantitate advantages and disadvantages of approaches. RECENT FINDINGS Overall, use of AI decreased the time spent to write the review article, but required more extensive fact checking. With the AI-only approach, up to 70% of the references cited were found to be inaccurate. Interestingly, the AI-assisted approach resulted in the highest similarity indices suggesting a higher likelihood of plagiarism. Finally, although the technology is rapidly changing, at the time of study, ChatGPT 4.0 had a cutoff date of September 2021 rendering identification of recent articles impossible. Therefore, all literature published past the cutoff date was manually provided to ChatGPT, rendering approaches #2 and #3 identical for contemporary citations. As a result, for the COVID-19 and musculoskeletal health topic, approach #2 was abandoned midstream due to the extensive overlap with approach #3. The main objective of this scientific study was to see whether AI could be used in a scientifically appropriate manner to improve the scientific writing process. Indeed, AI reduced the time for writing but had significant inaccuracies. The latter necessitates that AI cannot currently be used alone but could be used with careful oversight by humans to assist in writing scientific review articles.
Collapse
Affiliation(s)
- Melissa A Kacena
- Department of Orthopaedic Surgery, Indiana University School of Medicine, Indianapolis, IN, 46202, USA.
- Department of Anatomy, Cell Biology & Physiology, Indiana University School of Medicine, Indianapolis, IN, 46202, USA.
- Indiana Center for Musculoskeletal Health, Indiana University School of Medicine, Indianapolis, IN, 46202, USA.
- Richard L. Roudebush VA Medical Center, Indianapolis, IN, 46202, USA.
| | - Lilian I Plotkin
- Department of Anatomy, Cell Biology & Physiology, Indiana University School of Medicine, Indianapolis, IN, 46202, USA
- Indiana Center for Musculoskeletal Health, Indiana University School of Medicine, Indianapolis, IN, 46202, USA
- Richard L. Roudebush VA Medical Center, Indianapolis, IN, 46202, USA
| | - Jill C Fehrenbacher
- Indiana Center for Musculoskeletal Health, Indiana University School of Medicine, Indianapolis, IN, 46202, USA.
- Department of Pharmacology and Toxicology, Indiana University School of Medicine, Indianapolis, IN, 46202, USA.
- Stark Neuroscience Research Institute, Indiana University School of Medicine, Indianapolis, IN, 46202, USA.
| |
Collapse
|
105
|
Kapsali MZ, Livanis E, Tsalikidis C, Oikonomou P, Voultsos P, Tsaroucha A. Ethical Concerns About ChatGPT in Healthcare: A Useful Tool or the Tombstone of Original and Reflective Thinking? Cureus 2024; 16:e54759. [PMID: 38523987 PMCID: PMC10961144 DOI: 10.7759/cureus.54759] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 02/23/2024] [Indexed: 03/26/2024] Open
Abstract
Artificial intelligence (AI), the uprising technology of computer science aiming to create digital systems with human behavior and intelligence, seems to have invaded almost every field of modern life. Launched in November 2022, ChatGPT (Chat Generative Pre-trained Transformer) is a textual AI application capable of creating human-like responses characterized by original language and high coherence. Although AI-based language models have demonstrated impressive capabilities in healthcare, ChatGPT has received controversial annotations from the scientific and academic communities. This chatbot already appears to have a massive impact as an educational tool for healthcare professionals and transformative potential for clinical practice and could lead to dramatic changes in scientific research. Nevertheless, rational concerns were raised regarding whether the pre-trained, AI-generated text would be a menace not only for original thinking and new scientific ideas but also for academic and research integrity, as it gets more and more difficult to distinguish its AI origin due to the coherence and fluency of the produced text. This short review aims to summarize the potential applications and the consequential implications of ChatGPT in the three critical pillars of medicine: education, research, and clinical practice. In addition, this paper discusses whether the current use of this chatbot is in compliance with the ethical principles for the safe use of AI in healthcare, as determined by the World Health Organization. Finally, this review highlights the need for an updated ethical framework and the increased vigilance of healthcare stakeholders to harvest the potential benefits and limit the imminent dangers of this new innovative technology.
Collapse
Affiliation(s)
- Marina Z Kapsali
- Postgraduate Program on Bioethics, Laboratory of Bioethics, Democritus University of Thrace, Alexandroupolis, GRC
| | - Efstratios Livanis
- Department of Accounting and Finance, University of Macedonia, Thessaloniki, GRC
| | - Christos Tsalikidis
- Department of General Surgery, Democritus University of Thrace, Alexandroupolis, GRC
| | - Panagoula Oikonomou
- Laboratory of Experimental Surgery, Department of General Surgery, Democritus University of Thrace, Alexandroupolis, GRC
| | - Polychronis Voultsos
- Laboratory of Forensic Medicine & Toxicology (Medical Law and Ethics), School of Medicine, Faculty of Health Sciences, Aristotle University of Thessaloniki, Thessaloniki, GRC
| | - Aleka Tsaroucha
- Department of General Surgery, Democritus University of Thrace, Alexandroupolis, GRC
| |
Collapse
|
106
|
Perkins M, Roe J. Academic publisher guidelines on AI usage: A ChatGPT supported thematic analysis. F1000Res 2024; 12:1398. [PMID: 38322309 PMCID: PMC10844801 DOI: 10.12688/f1000research.142411.2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 01/10/2024] [Indexed: 02/08/2024] Open
Abstract
Background As Artificial Intelligence (AI) technologies such as Generative AI (GenAI) have become more common in academic settings, it is necessary to examine how these tools interact with issues of authorship, academic integrity, and research methodologies. The current landscape lacks cohesive policies and guidelines for regulating AI's role in academic research which has prompted discussions among publishers, authors, and institutions. Methods This study employs inductive thematic analysis to explore publisher policies regarding AI-assisted authorship and academic work. Our methods involved a two-fold analysis using both AI-assisted and traditional unassisted techniques to examine the available policies from leading academic publishers and other publishing or academic entities. The framework was designed to offer multiple perspectives, harnessing the strengths of AI for pattern recognition while leveraging human expertise for nuanced interpretation. The results of these two analyses are combined to form the final themes. Results Our findings indicate six overall themes, three of which were independently identified in both the AI-assisted and unassisted, manual analysis using common software tools. A broad consensus appears among publishers that human authorship remains paramount and that the use of GenAI tools is permissible but must be disclosed. However, GenAI tools are increasingly acknowledged for their supportive roles, including text generation and data analysis. The study also discusses the inherent limitations and biases of AI-assisted analysis, necessitating rigorous scrutiny by authors, reviewers, and editors. Conclusions There is a growing recognition of AI's role as a valuable auxiliary tool in academic research, but one that comes with caveats pertaining to integrity, accountability, and interpretive limitations. This study used a novel analysis supported by GenAI tools to identify themes emerging in the policy landscape, underscoring the need for an informed, flexible approach to policy formulation that can adapt to the rapidly evolving landscape of AI technologies.
Collapse
Affiliation(s)
| | - Jasper Roe
- James Cook University Singapore, Singapore, Singapore
| |
Collapse
|
107
|
Farhat F, Silva ES, Hassani H, Madsen DØ, Sohail SS, Himeur Y, Alam MA, Zafar A. The scholarly footprint of ChatGPT: a bibliometric analysis of the early outbreak phase. Front Artif Intell 2024; 6:1270749. [PMID: 38249789 PMCID: PMC10797012 DOI: 10.3389/frai.2023.1270749] [Citation(s) in RCA: 7] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Accepted: 12/08/2023] [Indexed: 01/23/2024] Open
Abstract
This paper presents a comprehensive analysis of the scholarly footprint of ChatGPT, an AI language model, using bibliometric and scientometric methods. The study zooms in on the early outbreak phase from when ChatGPT was launched in November 2022 to early June 2023. It aims to understand the evolution of research output, citation patterns, collaborative networks, application domains, and future research directions related to ChatGPT. By retrieving data from the Scopus database, 533 relevant articles were identified for analysis. The findings reveal the prominent publication venues, influential authors, and countries contributing to ChatGPT research. Collaborative networks among researchers and institutions are visualized, highlighting patterns of co-authorship. The application domains of ChatGPT, such as customer support and content generation, are examined. Moreover, the study identifies emerging keywords and potential research areas for future exploration. The methodology employed includes data extraction, bibliometric analysis using various indicators, and visualization techniques such as Sankey diagrams. The analysis provides valuable insights into ChatGPT's early footprint in academia and offers researchers guidance for further advancements. This study stimulates discussions, collaborations, and innovations to enhance ChatGPT's capabilities and impact across domains.
Collapse
Affiliation(s)
- Faiza Farhat
- Department of Zoology, Aligarh Muslim University, Aligarh, India
| | - Emmanuel Sirimal Silva
- Department of Economics and Law, Glasgow School for Business and Society, Glasgow Caledonian University, Glasgow, United Kingdom
| | - Hossein Hassani
- The Research Institute of Energy Management and Planning (RIEMP), University of Tehran, Tehran, Iran
| | - Dag Øivind Madsen
- USN School of Business, University of South-Eastern Norway, Hønefoss, Norway
| | - Shahab Saquib Sohail
- Department of Computer Science and Engineering, School of Engineering Sciences and Technology, Jamia Hamdard, New Delhi, India
| | - Yassine Himeur
- College of Engineering and Information Technology, University of Dubai, Dubai, United Arab Emirates
| | - M. Afshar Alam
- Department of Computer Science and Engineering, School of Engineering Sciences and Technology, Jamia Hamdard, New Delhi, India
| | - Aasim Zafar
- Department of Computer Science, Aligarh Muslim University, Aligarh, India
| |
Collapse
|
108
|
Khalifa AA, Ibrahim MA. Artificial intelligence (AI) and ChatGPT involvement in scientific and medical writing, a new concern for researchers. A scoping review. ARAB GULF JOURNAL OF SCIENTIFIC RESEARCH 2024. [DOI: 10.1108/agjsr-09-2023-0423] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/30/2024]
Abstract
PurposeThe study aims to evaluate PubMed publications on ChatGPT or artificial intelligence (AI) involvement in scientific or medical writing and investigate whether ChatGPT or AI was used to create these articles or listed as authors.Design/methodology/approachThis scoping review was conducted according to Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews (PRISMA-ScR) guidelines. A PubMed database search was performed for articles published between January 1 and November 29, 2023, using appropriate search terms; both authors performed screening and selection independently.FindingsFrom the initial search results of 127 articles, 41 were eligible for final analysis. Articles were published in 34 journals. Editorials were the most common article type, with 15 (36.6%) articles. Authors originated from 27 countries, and authors from the USA contributed the most, with 14 (34.1%) articles. The most discussed topic was AI tools and writing capabilities in 19 (46.3%) articles. AI or ChatGPT was involved in manuscript preparation in 31 (75.6%) articles. None of the articles listed AI or ChatGPT as an author, and in 19 (46.3%) articles, the authors acknowledged utilizing AI or ChatGPT.Practical implicationsResearchers worldwide are concerned with AI or ChatGPT involvement in scientific research, specifically the writing process. The authors believe that precise and mature regulations will be developed soon by journals, publishers and editors, which will pave the way for the best usage of these tools.Originality/valueThis scoping review expressed data published on using AI or ChatGPT in various scientific research and writing aspects, besides alluding to the advantages, disadvantages and implications of their usage.
Collapse
|
109
|
Muñoz-Cantero JM, Espiñeira-Bellón EM. Intelligent Plagiarism as a Misconduct in Academic Integrity. ACTA MEDICA PORT 2024; 37:1-2. [PMID: 38035394 DOI: 10.20344/amp.20233] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2023] [Accepted: 10/04/2023] [Indexed: 12/02/2023]
|
110
|
Guimarães NS, Joviano-Santos JV, Reis MG, Chaves RRM. Development of search strategies for systematic reviews in health using ChatGPT: a critical analysis. J Transl Med 2024; 22:1. [PMID: 38167166 PMCID: PMC10759630 DOI: 10.1186/s12967-023-04371-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2023] [Accepted: 07/17/2023] [Indexed: 01/05/2024] Open
Affiliation(s)
- Nathalia Sernizon Guimarães
- Postgraduate Health Science, Medical Sciences College of Minas Gerais, Alameda Ezequiel Dias, 275, Belo Horizonte, Minas Gerais, 30130-110, Brazil.
| | - Julliane Vasconcelos Joviano-Santos
- Postgraduate Health Science, Medical Sciences College of Minas Gerais, Alameda Ezequiel Dias, 275, Belo Horizonte, Minas Gerais, 30130-110, Brazil
| | - Marcela Gomes Reis
- Postgraduate Health Science, Medical Sciences College of Minas Gerais, Alameda Ezequiel Dias, 275, Belo Horizonte, Minas Gerais, 30130-110, Brazil
| | - Roberta Rayra Martins Chaves
- Postgraduate Health Science, Medical Sciences College of Minas Gerais, Alameda Ezequiel Dias, 275, Belo Horizonte, Minas Gerais, 30130-110, Brazil
| |
Collapse
|
111
|
Zybaczynska J, Norris M, Modi S, Brennan J, Jhaveri P, Craig TJ, Al-Shaikhly T. Artificial Intelligence-Generated Scientific Literature: A Critical Appraisal. THE JOURNAL OF ALLERGY AND CLINICAL IMMUNOLOGY. IN PRACTICE 2024; 12:106-110. [PMID: 37832818 DOI: 10.1016/j.jaip.2023.10.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/25/2023] [Revised: 08/14/2023] [Accepted: 10/03/2023] [Indexed: 10/15/2023]
Abstract
BACKGROUND Review articles play a critical role in informing medical decisions and identifying avenues for future research. With the introduction of artificial intelligence (AI), there has been a growing interest in the potential of this technology to transform the synthesis of medical literature. Open AI's Generative Pre-trained Transformer (GPT-4) (Open AI Inc, San Francisco, CA) tool provides access to advanced AI that is able to quickly produce medical literature following only simple prompts. The accuracy of the generated articles requires review, especially in subspecialty fields like Allergy/Immunology. OBJECTIVE To critically appraise AI-synthesized allergy-focused minireviews. METHODS We tasked the GPT-4 Chatbot with generating 2 1,000-word reviews on the topics of hereditary angioedema and eosinophilic esophagitis. Authors critically appraised these articles using the Joanna Briggs Institute (JBI) tool for text and opinion and additionally evaluated domains of interest such as language, reference quality, and accuracy of the content. RESULTS The language of the AI-generated minireviews was carefully articulated and logically focused on the topic of interest; however, reviewers of the AI-generated articles indicated that the AI-generated content lacked depth, did not appear to be the result of an analytical process, missed critical information, and contained inaccurate information. Despite being provided instruction to utilize scientific references, the AI chatbot relied mainly on freely available resources, and the AI chatbot fabricated references. CONCLUSIONS The AI holds the potential to change the landscape of synthesizing medical literature; however, apparent inaccurate and fabricated information calls for rigorous evaluation and validation of AI tools in generating medical literature, especially on subjects associated with limited resources.
Collapse
Affiliation(s)
- Justyna Zybaczynska
- Section of Allergy, Asthma & Immunology, Department of Medicine, Pennsylvania State University College of Medicine, Hershey, Pa
| | - Matthew Norris
- Section of Allergy, Asthma & Immunology, Department of Medicine, Pennsylvania State University College of Medicine, Hershey, Pa
| | - Sunjay Modi
- Section of Allergy, Asthma & Immunology, Department of Medicine, Pennsylvania State University College of Medicine, Hershey, Pa
| | - Jennifer Brennan
- Section of Allergy, Asthma & Immunology, Department of Medicine, Pennsylvania State University College of Medicine, Hershey, Pa
| | - Pooja Jhaveri
- Division of Allergy & Immunology, Department of Pediatrics, Pennsylvania State University College of Medicine, Hershey, Pa
| | - Timothy J Craig
- Section of Allergy, Asthma & Immunology, Department of Medicine, Pennsylvania State University College of Medicine, Hershey, Pa
| | - Taha Al-Shaikhly
- Section of Allergy, Asthma & Immunology, Department of Medicine, Pennsylvania State University College of Medicine, Hershey, Pa.
| |
Collapse
|
112
|
Hu X, Niemann M, Kienzle A, Braun K, Back DA, Gwinner C, Renz N, Stoeckle U, Trampuz A, Meller S. Evaluating ChatGPT responses to frequently asked patient questions regarding periprosthetic joint infection after total hip and knee arthroplasty. Digit Health 2024; 10:20552076241272620. [PMID: 39130521 PMCID: PMC11311159 DOI: 10.1177/20552076241272620] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2024] [Accepted: 07/09/2024] [Indexed: 08/13/2024] Open
Abstract
Background Patients access relevant information concerning their orthopaedic surgery resources through multiple information channels before presenting for clinical treatment. Recently, artificial intelligence (AI)-powered chatbots have become another source of information for patients. The currently developed AI chat technology ChatGPT (OpenAI LP) is an application for such purposes and it has been rapidly gaining popularity, including for patient education. This study sought to evaluate whether ChatGPT can correctly answer frequently asked questions (FAQ) regarding periprosthetic joint infection (PJI). Methods Twelve FAQs about PJI after hip and knee arthroplasty were identified from the websites of fifteen international clinical expert centres. ChatGPT was confronted with these questions and its responses were analysed for their accuracy using an evidence-based approach by a multidisciplinary team. Responses were categorised in four groups: (1) Excellent response that did not require additional improvement; (2) Satisfactory responses that required a small amount of improvement; (3) Satisfactory responses that required moderate improvement; and (4) Unsatisfactory responses that required a large amount of improvement. Results From the analysis of the responses given by the chatbot, no reply received an 'unsatisfactory' rating; one did not require any correction; and the majority of the responses required low (7 out of 12) or moderate (4 out of 12) clarification. Although a few responses required minimal clarification, the chatbot responses were generally unbiased and evidence-based, even when asked controversial questions. Conclusions The AI-chatbot ChatGPT was able to effectively answer the FAQs of patients seeking information around PJI diagnosis and treatment. The given information was also written in a manner that can be assumed to be understandable by patients. The chatbot could be a valuable clinical tool for patient education and understanding around PJI treatment in the future. Further studies should evaluate its use and acceptance by patients with PJI.
Collapse
Affiliation(s)
- Xiaojun Hu
- Center for Musculoskeletal Surgery, Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin, Humboldt-Universität zu Berlin, and Berlin Institute of Health, Berlin, Germany
- Department of Orthopedics, Seventh People's Hospital of Chongqing, Chongqing, China
| | - Marcel Niemann
- Center for Musculoskeletal Surgery, Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin, Humboldt-Universität zu Berlin, and Berlin Institute of Health, Berlin, Germany
| | - Arne Kienzle
- Center for Musculoskeletal Surgery, Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin, Humboldt-Universität zu Berlin, and Berlin Institute of Health, Berlin, Germany
| | - Karl Braun
- Department of Trauma Surgery, University Hospital Rechts der Isar, Technical University of Munich, Munich, Germany
| | - David Alexander Back
- Center for Musculoskeletal Surgery, Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin, Humboldt-Universität zu Berlin, and Berlin Institute of Health, Berlin, Germany
| | - Clemens Gwinner
- Center for Musculoskeletal Surgery, Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin, Humboldt-Universität zu Berlin, and Berlin Institute of Health, Berlin, Germany
| | - Nora Renz
- Center for Musculoskeletal Surgery, Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin, Humboldt-Universität zu Berlin, and Berlin Institute of Health, Berlin, Germany
| | - Ulrich Stoeckle
- Center for Musculoskeletal Surgery, Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin, Humboldt-Universität zu Berlin, and Berlin Institute of Health, Berlin, Germany
| | - Andrej Trampuz
- Center for Musculoskeletal Surgery, Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin, Humboldt-Universität zu Berlin, and Berlin Institute of Health, Berlin, Germany
| | - Sebastian Meller
- Center for Musculoskeletal Surgery, Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin, Humboldt-Universität zu Berlin, and Berlin Institute of Health, Berlin, Germany
| |
Collapse
|
113
|
Sumbal A, Sumbal R, Amir A. Can ChatGPT-3.5 Pass a Medical Exam? A Systematic Review of ChatGPT's Performance in Academic Testing. JOURNAL OF MEDICAL EDUCATION AND CURRICULAR DEVELOPMENT 2024; 11:23821205241238641. [PMID: 38487300 PMCID: PMC10938614 DOI: 10.1177/23821205241238641] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Accepted: 02/25/2024] [Indexed: 03/17/2024]
Abstract
OBJECTIVE We, therefore, aim to conduct a systematic review to assess the academic potential of ChatGPT-3.5, along with its strengths and limitations when giving medical exams. METHOD Following PRISMA guidelines, a systemic search of the literature was performed using electronic databases PUBMED/MEDLINE, Google Scholar, and Cochrane. Articles from their inception till April 4, 2023, were queried. A formal narrative analysis was conducted by systematically arranging similarities and differences between individual findings together. RESULTS After rigorous screening, 12 articles underwent this review. All the selected papers assessed the academic performance of ChatGPT-3.5. One study compared the performance of ChatGPT-3.5 with the performance of ChatGPT-4 when giving a medical exam. Overall, ChatGPT performed well in 4 tests, averaged in 4 tests, and performed badly in 4 tests. ChatGPT's performance was directly proportional to the level of the questions' difficulty but was unremarkable on whether the questions were binary, descriptive, or MCQ-based. ChatGPT's explanation, reasoning, memory, and accuracy were remarkably good, whereas it failed to understand image-based questions, and lacked insight and critical thinking. CONCLUSION ChatGPT-3.5 performed satisfactorily in the exams it took as an examinee. However, there is a need for future related studies to fully explore the potential of ChatGPT in medical education.
Collapse
Affiliation(s)
- Anusha Sumbal
- Dow University of Health Sciences, Karachi, Pakistan
| | - Ramish Sumbal
- Dow University of Health Sciences, Karachi, Pakistan
| | - Alina Amir
- Dow University of Health Sciences, Karachi, Pakistan
| |
Collapse
|
114
|
Thibaut G, Dabbagh A, Liverneaux P. Does Google's Bard Chatbot perform better than ChatGPT on the European hand surgery exam? INTERNATIONAL ORTHOPAEDICS 2024; 48:151-158. [PMID: 37968408 DOI: 10.1007/s00264-023-06034-y] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/07/2023] [Accepted: 11/01/2023] [Indexed: 11/17/2023]
Abstract
PURPOSE According to a previous research, the chatbot ChatGPT® V3.5 was unable to pass the first part of the European Board of Hand Surgery (EBHS) diploma examination. This study aimed to investigate whether Google's chatbot Bard® would have superior performance compared to ChatGPT on the EBHS diploma examination. METHODS Chatbots were asked to answer 18 EBHS multiple choice questions (MCQs) published in the Journal of Hand Surgery (European Volume) in five trials (A1 to A5). After A3, chatbots received correct answers, and after A4, incorrect answers. Consequently, their ability to modify their response was measured and compared. RESULTS Bard® scored 3/18 (A1), 1/18 (A2), 4/18 (A3) and 2/18 (A4 and A5). The average percentage of correct answers was 61.1% for A1, 62.2% for A2, 64.4% for A3, 65.6% for A4, 63.3% for A5 and 63.3% for all trials combined. Agreement was moderate from A1 to A5 (kappa = 0.62 (IC95% = [0.51; 0.73])) as well as from A1 to A3 (kappa = 0.60 (IC95% = [0.47; 0.74])). The formulation of Bard® responses was homogeneous, but its learning capacity is still developing. CONCLUSIONS The main hypothesis of our study was not proved since Bard did not score significantly higher than ChatGPT when answering the MCQs of the EBHS diploma exam. In conclusion, neither ChatGPT® nor Bard®, in their current versions, can pass the first part of the EBHS diploma exam.
Collapse
Affiliation(s)
- Goetsch Thibaut
- Department of Public Health, Strasbourg University Hospital, FMTS, GMRC, 1 avenue de l'hôpital, 67000, Strasbourg cedex, France
| | | | - Philippe Liverneaux
- ICube CNRS UMR7357, Strasbourg University, 2-4 rue Boussingault, 67000, Strasbourg, France.
- Department of Hand Surgery, Strasbourg University Hospitals, FMTS, 1 avenue Molière, 67200, Strasbourg, France.
| |
Collapse
|
115
|
Talyshinskii A, Naik N, Hameed BMZ, Juliebø-Jones P, Somani BK. Potential of AI-Driven Chatbots in Urology: Revolutionizing Patient Care Through Artificial Intelligence. Curr Urol Rep 2024; 25:9-18. [PMID: 37723300 PMCID: PMC10787686 DOI: 10.1007/s11934-023-01184-3] [Citation(s) in RCA: 12] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/05/2023] [Indexed: 09/20/2023]
Abstract
PURPOSE OF REVIEW Artificial intelligence (AI) chatbots have emerged as a potential tool to transform urology by improving patient care and physician efficiency. With an emphasis on their potential advantages and drawbacks, this literature review offers a thorough assessment of the state of AI-driven chatbots in urology today. RECENT FINDINGS The capacity of AI-driven chatbots in urology to give patients individualized and timely medical advice is one of its key advantages. Chatbots can help patients prioritize their symptoms and give advice on the best course of treatment. By automating administrative duties and offering clinical decision support, chatbots can also help healthcare providers. Before chatbots are widely used in urology, there are a few issues that need to be resolved. The precision of chatbot diagnoses and recommendations might be impacted by technical constraints like system errors and flaws. Additionally, issues regarding the security and privacy of patient data must be resolved, and chatbots must adhere to all applicable laws. Important issues that must be addressed include accuracy and dependability because any mistakes or inaccuracies could seriously harm patients. The final obstacle is resistance from patients and healthcare professionals who are hesitant to use new technology or who value in-person encounters. AI-driven chatbots have the potential to significantly improve urology care and efficiency. However, it is essential to thoroughly test and ensure the accuracy of chatbots, address privacy and security concerns, and design user-friendly chatbots that can integrate into existing workflows. By exploring various scenarios and examining the current literature, this review provides an analysis of the prospects and limitations of implementing chatbots in urology.
Collapse
Affiliation(s)
- Ali Talyshinskii
- Department of Urology, Astana Medical University, Astana, Kazakhstan
| | - Nithesh Naik
- Department of Mechanical and Industrial Engineering, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - B M Zeeshan Hameed
- Department of Urology, Father Muller Medical College, Mangalore, Karnataka, India
| | - Patrick Juliebø-Jones
- Department of Urology, Haukeland University Hospital, Bergen, Norway.
- Department of Clinical Medicine, University of Bergen, Bergen, Norway.
| | | |
Collapse
|
116
|
Lechien JR, Georgescu BM, Hans S, Chiesa-Estomba CM. ChatGPT performance in laryngology and head and neck surgery: a clinical case-series. Eur Arch Otorhinolaryngol 2024; 281:319-333. [PMID: 37874336 DOI: 10.1007/s00405-023-08282-5] [Citation(s) in RCA: 24] [Impact Index Per Article: 24.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2023] [Accepted: 10/06/2023] [Indexed: 10/25/2023]
Abstract
OBJECTIVES To study the performance of ChatGPT in the management of laryngology and head and neck (LHN) cases. METHODS History and clinical examination of patients consulting at the Otolaryngology-Head and Neck Surgery department were presented to ChatGPT, which was interrogated for differential diagnosis, management, and treatment. The ChatGPT performance was assessed by two blinded board-certified otolaryngologists using the following items of a composite score and the Ottawa Clinic Assessment Tool: differential diagnosis; additional examination; and treatment options. The complexity of clinical cases was evaluated with the Amsterdam Clinical Challenge Scale test. RESULTS Forty clinical cases were submitted to ChatGPT, accounting for 14 (35%), 12 (30%), and 14 (35%) easy, moderate and difficult cases, respectively. ChatGPT indicated a significant higher number of additional examinations compared to practitioners (p = 0.001). There was a significant agreement between practitioners and ChatGPT for the indication of some common examinations (audiometry, ultrasonography, biopsy, gastrointestinal endoscopy or videofluoroscopy). ChatGPT never indicated some important additional examinations (PET-CT, voice quality assessment, or impedance-pH monitoring). ChatGPT reported highest performance in the proposition of the primary (90%) or the most plausible differential diagnoses (65%), and the therapeutic options (60-68%). The ChatGPT performance in the indication of additional examinations was lowest. CONCLUSIONS ChatGPT is a promising adjunctive tool in LHN practice, providing extensive documentation about disease-related additional examinations, differential diagnoses, and treatments. The ChatGPT is more efficient in diagnosis and treatment, rather than in the selection of the most adequate additional examination.
Collapse
Affiliation(s)
- Jerome R Lechien
- Research Committee of Young Otolaryngologists of the International Federation of Otorhinolaryngological Socities (IFOS), Paris, France.
- Division of Laryngology and Broncho-Esophagology, Department of Otolaryngology-Head Neck Surgery, UMONS Research Institute for Health Sciences and Technology, EpiCURA Hospital, University of Mons (UMons), Mons, Belgium.
- Department of Otorhinolaryngology and Head and Neck Surgery, School of Medicine, UFR Simone Veil, Foch Hospital, Université Versailles Saint-Quentin-en-Yvelines (Paris Saclay University), Paris, France.
- Department of Otorhinolaryngology and Head and Neck Surgery, CHU Saint-Pierre, Brussels, Belgium.
- Polyclinique Elsan de Poitiers, Poitiers, France.
- Department of Human Anatomy and Experimental Oncology, Faculty of Medicine, UMONS Research Institute for Health Sciences and Technology, Avenue du Champ de Mars, 6, 7000, Mons, Belgium.
| | - Bianca M Georgescu
- Division of Laryngology and Broncho-Esophagology, Department of Otolaryngology-Head Neck Surgery, UMONS Research Institute for Health Sciences and Technology, EpiCURA Hospital, University of Mons (UMons), Mons, Belgium
| | - Stephane Hans
- Research Committee of Young Otolaryngologists of the International Federation of Otorhinolaryngological Socities (IFOS), Paris, France
- Department of Otorhinolaryngology and Head and Neck Surgery, School of Medicine, UFR Simone Veil, Foch Hospital, Université Versailles Saint-Quentin-en-Yvelines (Paris Saclay University), Paris, France
| | - Carlos M Chiesa-Estomba
- Research Committee of Young Otolaryngologists of the International Federation of Otorhinolaryngological Socities (IFOS), Paris, France
- Department of Otorhinolaryngology-Head & Neck Surgery, Donostia University Hospital-Biodonostia Research Institute, St. Sebastian, Spain
| |
Collapse
|
117
|
Sogi GM. Decipher the Cipher. Contemp Clin Dent 2024; 15:1-2. [PMID: 38707670 PMCID: PMC11068239 DOI: 10.4103/ccd.ccd_112_24] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/07/2024] Open
Affiliation(s)
- Girish Malleshappa Sogi
- Editor-in-chief, Contemporary Clinical Dentistry, Principal cum Dean, MM College of Dental Sciences and Research, Maharishi Markandeshwar (Deemed to be University), Ambala, Haryana, India E-mail:
| |
Collapse
|
118
|
BaHammam AS. Balancing Innovation and Integrity: The Role of AI in Research and Scientific Writing. Nat Sci Sleep 2023; 15:1153-1156. [PMID: 38170140 PMCID: PMC10759812 DOI: 10.2147/nss.s455765] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/19/2023] [Accepted: 12/21/2023] [Indexed: 01/05/2024] Open
Affiliation(s)
- Ahmed S BaHammam
- Editor-in-Chief Nature and Science of Sleep
- Department of Medicine, University Sleep Disorders Center and Pulmonary Service, King Saud University, Riyadh, Saudi Arabia
- King Saud University Medical City, Riyadh, Saudi Arabia
| |
Collapse
|
119
|
Cheng SL, Tsai SJ, Bai YM, Ko CH, Hsu CW, Yang FC, Tsai CK, Tu YK, Yang SN, Tseng PT, Hsu TW, Liang CS, Su KP. Comparisons of Quality, Correctness, and Similarity Between ChatGPT-Generated and Human-Written Abstracts for Basic Research: Cross-Sectional Study. J Med Internet Res 2023; 25:e51229. [PMID: 38145486 PMCID: PMC10760418 DOI: 10.2196/51229] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2023] [Revised: 10/17/2023] [Accepted: 11/20/2023] [Indexed: 12/26/2023] Open
Abstract
BACKGROUND ChatGPT may act as a research assistant to help organize the direction of thinking and summarize research findings. However, few studies have examined the quality, similarity (abstracts being similar to the original one), and accuracy of the abstracts generated by ChatGPT when researchers provide full-text basic research papers. OBJECTIVE We aimed to assess the applicability of an artificial intelligence (AI) model in generating abstracts for basic preclinical research. METHODS We selected 30 basic research papers from Nature, Genome Biology, and Biological Psychiatry. Excluding abstracts, we inputted the full text into ChatPDF, an application of a language model based on ChatGPT, and we prompted it to generate abstracts with the same style as used in the original papers. A total of 8 experts were invited to evaluate the quality of these abstracts (based on a Likert scale of 0-10) and identify which abstracts were generated by ChatPDF, using a blind approach. These abstracts were also evaluated for their similarity to the original abstracts and the accuracy of the AI content. RESULTS The quality of ChatGPT-generated abstracts was lower than that of the actual abstracts (10-point Likert scale: mean 4.72, SD 2.09 vs mean 8.09, SD 1.03; P<.001). The difference in quality was significant in the unstructured format (mean difference -4.33; 95% CI -4.79 to -3.86; P<.001) but minimal in the 4-subheading structured format (mean difference -2.33; 95% CI -2.79 to -1.86). Among the 30 ChatGPT-generated abstracts, 3 showed wrong conclusions, and 10 were identified as AI content. The mean percentage of similarity between the original and the generated abstracts was not high (2.10%-4.40%). The blinded reviewers achieved a 93% (224/240) accuracy rate in guessing which abstracts were written using ChatGPT. CONCLUSIONS Using ChatGPT to generate a scientific abstract may not lead to issues of similarity when using real full texts written by humans. However, the quality of the ChatGPT-generated abstracts was suboptimal, and their accuracy was not 100%.
Collapse
Affiliation(s)
- Shu-Li Cheng
- Department of Nursing, Mackay Medical College, Taipei, Taiwan
| | - Shih-Jen Tsai
- Department of Psychiatry, Taipei Veterans General Hospital, Taipei, Taiwan
- Division of Psychiatry, School of Medicine, National Yang-Ming University, Taipei, Taiwan
| | - Ya-Mei Bai
- Department of Psychiatry, Taipei Veterans General Hospital, Taipei, Taiwan
- Division of Psychiatry, School of Medicine, National Yang-Ming University, Taipei, Taiwan
| | - Chih-Hung Ko
- Department of Psychiatry, Kaohsiung Medical University Hospital, Kaohsiung, Taiwan
- Department of Psychiatry, College of Medicine, Kaohsiung Medical University, Kaohsiung, Taiwan
- Department of Psychiatry, Kaohsiung Municipal Siaogang Hospital, Kaohsiung Medical University, Kaohsiung, Taiwan
| | - Chih-Wei Hsu
- Department of Psychiatry, Kaohsiung Chang Gung Memorial Hospital, Kaohsiung, Taiwan
| | - Fu-Chi Yang
- Department of Neurology, Tri-Service General Hospital, National Defense Medical Center, Taipei, Taiwan
| | - Chia-Kuang Tsai
- Department of Neurology, Tri-Service General Hospital, National Defense Medical Center, Taipei, Taiwan
| | - Yu-Kang Tu
- Institute of Epidemiology and Preventive Medicine, College of Public Health, National Taiwan University, Taipei, Taiwan
- Department of Dentistry, National Taiwan University Hospital, Taipei, Taiwan
| | - Szu-Nian Yang
- Department of Psychiatry, Tri-service Hospital, Beitou branch, Taipei, Taiwan
- Department of Psychiatry, Armed Forces Taoyuan General Hospital, Taoyuan, Taiwan
- Graduate Institute of Health and Welfare Policy, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Ping-Tao Tseng
- Institute of Biomedical Sciences, Institute of Precision Medicine, National Sun Yat-sen University, Kaohsiung, Taiwan
- Department of Psychology, College of Medical and Health Science, Asia University, Taichung, Taiwan
- Prospect Clinic for Otorhinolaryngology and Neurology, Kaohsiung, Taiwan
| | - Tien-Wei Hsu
- Department of Psychiatry, E-Da Dachang Hospital, I-Shou University, Kaohsiung, Taiwan
- Department of Psychiatry, E-Da Hospital, I-Shou University, Kaohsiung, Taiwan
| | - Chih-Sung Liang
- Department of Psychiatry, Tri-service Hospital, Beitou branch, Taipei, Taiwan
- Department of Psychiatry, National Defense Medical Center, Taipei, Taiwan
| | - Kuan-Pin Su
- College of Medicine, China Medical University, Taichung, Taiwan
- Mind-Body Interface Laboratory, China Medical University and Hospital, Taichung, Taiwan
- An-Nan Hospital, China Medical University, Tainan, Taiwan
| |
Collapse
|
120
|
Ferreira RM. New evidence-based practice: Artificial intelligence as a barrier breaker. World J Methodol 2023; 13:384-389. [PMID: 38229944 PMCID: PMC10789101 DOI: 10.5662/wjm.v13.i5.384] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/17/2023] [Revised: 10/24/2023] [Accepted: 11/08/2023] [Indexed: 12/20/2023] Open
Abstract
The concept of evidence-based practice has persisted over several years and remains a cornerstone in clinical practice, representing the gold standard for optimal patient care. However, despite widespread recognition of its significance, practical application faces various challenges and barriers, including a lack of skills in interpreting studies, limited resources, time constraints, linguistic competencies, and more. Recently, we have witnessed the emergence of a groundbreaking technological revolution known as artificial intelligence. Although artificial intelligence has become increasingly integrated into our daily lives, some reluctance persists among certain segments of the public. This article explores the potential of artificial intelligence as a solution to some of the main barriers encountered in the application of evidence-based practice. It highlights how artificial intelligence can assist in staying updated with the latest evidence, enhancing clinical decision-making, addressing patient misinformation, and mitigating time constraints in clinical practice. The integration of artificial intelligence into evidence-based practice has the potential to revolutionize healthcare, leading to more precise diagnoses, personalized treatment plans, and improved doctor-patient interactions. This proposed synergy between evidence-based practice and artificial intelligence may necessitate adjustments to its core concept, heralding a new era in healthcare.
Collapse
Affiliation(s)
- Ricardo Maia Ferreira
- Department of Sports and Exercise, Polytechnic Institute of Maia (N2i), Maia 4475-690, Porto, Portugal
- Department of Physioterapy, Polytechnic Institute of Coimbra, Coimbra Health School, Coimbra 3046-854, Coimbra, Portugal
- Department of Physioterapy, Polytechnic Institute of Castelo Branco, Dr. Lopes Dias Health School, Castelo Branco 6000-767, Castelo Branco, Portugal
- Sport Physical Activity and Health Research & Innovation Center, Polytechnic Institute of Viana do Castelo, Melgaço, 4960-320, Viana do Castelo, Portugal
| |
Collapse
|
121
|
Song C, Song Y. Enhancing academic writing skills and motivation: assessing the efficacy of ChatGPT in AI-assisted language learning for EFL students. Front Psychol 2023; 14:1260843. [PMID: 38162975 PMCID: PMC10754989 DOI: 10.3389/fpsyg.2023.1260843] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2023] [Accepted: 11/28/2023] [Indexed: 01/03/2024] Open
Abstract
Introduction This mixed-methods study evaluates the impact of AI-assisted language learning on Chinese English as a Foreign Language (EFL) students' writing skills and writing motivation. As artificial intelligence (AI) becomes more prevalent in educational settings, understanding its effects on language learning outcomes is crucial. Methods The study employs a comprehensive approach, combining quantitative and qualitative methods. The quantitative phase utilizes a pre-test and post-test design to assess writing skills. Fifty EFL students, matched for proficiency, are randomly assigned to experimental (AI-assisted instruction via ChatGPT) or control (traditional instruction) groups. Writing samples are evaluated using established scoring rubrics. Concurrently, semi-structured interviews are conducted with a subset of participants to explore writing motivation and experiences with AI-assisted learning. Results Quantitative analysis reveals significant improvements in both writing skills and motivation among students who received AI-assisted instruction compared to the control group. The experimental group demonstrates enhanced proficiency in various aspects of writing, including organization, coherence, grammar, and vocabulary. Qualitative findings showcase diverse perspectives, ranging from recognition of AI's innovative instructional role and its positive influence on writing skills and motivation to concerns about contextual accuracy and over-reliance. Participants also reflect on the long-term impact and sustainability of AI-assisted instruction, emphasizing the need for ongoing development and adaptation of AI tools. Discussion The nuanced findings offer a comprehensive understanding of AI's transformative potential in education. These insights have practical implications for practitioners and researchers, emphasizing the benefits, challenges, and the evolving nature of AI's role in language instruction.
Collapse
Affiliation(s)
- Cuiping Song
- School of Foreign Studies, North Minzu University, Yinchuan, Ningxia, China
| | - Yanping Song
- School of Public Administration, Central South University, Changsha, Hunan, China
| |
Collapse
|
122
|
Semrl N, Feigl S, Taumberger N, Bracic T, Fluhr H, Blockeel C, Kollmann M. AI language models in human reproduction research: exploring ChatGPT's potential to assist academic writing. Hum Reprod 2023; 38:2281-2288. [PMID: 37833847 DOI: 10.1093/humrep/dead207] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2023] [Revised: 09/06/2023] [Indexed: 10/15/2023] Open
Abstract
Artificial intelligence (AI)-driven language models have the potential to serve as an educational tool, facilitate clinical decision-making, and support research and academic writing. The benefits of their use are yet to be evaluated and concerns have been raised regarding the accuracy, transparency, and ethical implications of using this AI technology in academic publishing. At the moment, Chat Generative Pre-trained Transformer (ChatGPT) is one of the most powerful and widely debated AI language models. Here, we discuss its feasibility to answer scientific questions, identify relevant literature, and assist writing in the field of human reproduction. With consideration of the scarcity of data on this topic, we assessed the feasibility of ChatGPT in academic writing, using data from six meta-analyses published in a leading journal of human reproduction. The text generated by ChatGPT was evaluated and compared to the original text by blinded reviewers. While ChatGPT can produce high-quality text and summarize information efficiently, its current ability to interpret data and answer scientific questions is limited, and it cannot be relied upon for a literature search or accurate source citation due to the potential spread of incomplete or false information. We advocate for open discussions within the reproductive medicine research community to explore the advantages and disadvantages of implementing this AI technology. Researchers and reviewers should be informed about AI language models, and we encourage authors to transparently disclose their use.
Collapse
Affiliation(s)
- N Semrl
- Department of Obstetrics and Gynecology, Medical University of Graz, Graz, Austria
| | - S Feigl
- Department of Obstetrics and Gynecology, Medical University of Graz, Graz, Austria
| | - N Taumberger
- Department of Obstetrics and Gynecology, Medical University of Graz, Graz, Austria
| | - T Bracic
- Department of Obstetrics and Gynecology, Medical University of Graz, Graz, Austria
| | - H Fluhr
- Department of Obstetrics and Gynecology, Medical University of Graz, Graz, Austria
| | - C Blockeel
- Centre for Reproductive Medicine, Universitair Ziekenhuis Brussel (UZ Brussel), Brussels, Belgium
| | - M Kollmann
- Department of Obstetrics and Gynecology, Medical University of Graz, Graz, Austria
| |
Collapse
|
123
|
Miao J, Thongprayoon C, Suppadungsuk S, Garcia Valencia OA, Qureshi F, Cheungpasitporn W. Innovating Personalized Nephrology Care: Exploring the Potential Utilization of ChatGPT. J Pers Med 2023; 13:1681. [PMID: 38138908 PMCID: PMC10744377 DOI: 10.3390/jpm13121681] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2023] [Revised: 12/02/2023] [Accepted: 12/02/2023] [Indexed: 12/24/2023] Open
Abstract
The rapid advancement of artificial intelligence (AI) technologies, particularly machine learning, has brought substantial progress to the field of nephrology, enabling significant improvements in the management of kidney diseases. ChatGPT, a revolutionary language model developed by OpenAI, is a versatile AI model designed to engage in meaningful and informative conversations. Its applications in healthcare have been notable, with demonstrated proficiency in various medical knowledge assessments. However, ChatGPT's performance varies across different medical subfields, posing challenges in nephrology-related queries. At present, comprehensive reviews regarding ChatGPT's potential applications in nephrology remain lacking despite the surge of interest in its role in various domains. This article seeks to fill this gap by presenting an overview of the integration of ChatGPT in nephrology. It discusses the potential benefits of ChatGPT in nephrology, encompassing dataset management, diagnostics, treatment planning, and patient communication and education, as well as medical research and education. It also explores ethical and legal concerns regarding the utilization of AI in medical practice. The continuous development of AI models like ChatGPT holds promise for the healthcare realm but also underscores the necessity of thorough evaluation and validation before implementing AI in real-world medical scenarios. This review serves as a valuable resource for nephrologists and healthcare professionals interested in fully utilizing the potential of AI in innovating personalized nephrology care.
Collapse
Affiliation(s)
- Jing Miao
- Division of Nephrology and Hypertension, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA; (J.M.); (C.T.); (S.S.); (O.A.G.V.); (F.Q.)
| | - Charat Thongprayoon
- Division of Nephrology and Hypertension, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA; (J.M.); (C.T.); (S.S.); (O.A.G.V.); (F.Q.)
| | - Supawadee Suppadungsuk
- Division of Nephrology and Hypertension, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA; (J.M.); (C.T.); (S.S.); (O.A.G.V.); (F.Q.)
- Chakri Naruebodindra Medical Institute, Faculty of Medicine Ramathibodi Hospital, Mahidol University, Samut Prakan 10540, Thailand
| | - Oscar A. Garcia Valencia
- Division of Nephrology and Hypertension, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA; (J.M.); (C.T.); (S.S.); (O.A.G.V.); (F.Q.)
| | - Fawad Qureshi
- Division of Nephrology and Hypertension, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA; (J.M.); (C.T.); (S.S.); (O.A.G.V.); (F.Q.)
| | - Wisit Cheungpasitporn
- Division of Nephrology and Hypertension, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA; (J.M.); (C.T.); (S.S.); (O.A.G.V.); (F.Q.)
| |
Collapse
|
124
|
Wu Z, Bian L, Geng H, Zheng Z, Wang H, Zhai B. Application and challenges of ChatGPT in interventional surgery. Int J Surg 2023; 109:3747-3749. [PMID: 37713498 PMCID: PMC10720859 DOI: 10.1097/js9.0000000000000704] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2023] [Accepted: 08/12/2023] [Indexed: 09/17/2023]
Affiliation(s)
| | | | - Haigang Geng
- Department of Gastrointestinal Surgery, Renji Hospital, Shanghai Jiao Tong University School of Medicine
| | - Zhigang Zheng
- Department of Liver Surgery and Liver Transplantation, School of Medicine, Renji Hospital, Shanghai Jiao Tong University, Shanghai, People’s Republic of China
| | | | - Bo Zhai
- Department of Interventional Oncology
| |
Collapse
|
125
|
Gerli AG, Soriano JB, Alicandro G, Salvagno M, Taccone F, Centanni S, LA Vecchia C. ChatGPT: unlocking the potential of Artifical Intelligence in COVID-19 monitoring and prediction. Panminerva Med 2023; 65:461-466. [PMID: 37535043 DOI: 10.23736/s0031-0808.23.04853-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/04/2023]
Abstract
BACKGROUND The COVID-19 pandemic has had an unprecedent impact of everyday life with deleterious consequences on global health, economics, and society. Thus, accurate and timely information is critical for monitoring its spread and mitigating its impact. ChatGPT is a large language model chatbot with artificial intelligence, developed by OpenAI, that can provide both textual content and R code for predictive models. It may prove to be useful in analyzing and interpreting COVID-19-related data. METHODS This paper explores the application of ChatGPT to the monitoring of the COVID-19 pandemic, presenting R code for predictive models and demonstrating the model's capabilities in sentiment analysis, information extraction, and predictive modelling. We used the prediction models suggested by ChatGPT to predict the daily number of COVID-19 deaths in Italy. The prediction accuracy of the models was compared using the following metrics: mean squared error (MSE), mean absolute deviation (MAD) and root mean squared error (RMSE). RESULTS ChatGPT suggested three different predictive models, including ARIMA, Random Forest and Prophet. The ARIMA model outperformed the other two models in predicting the daily number of COVID-19 deaths in Italy, with lower MSE, MAD, and RMSE values as compared to the Random Forest and Prophet. CONCLUSIONS This paper demonstrates the potential of ChatGPT as a valuable tool in the monitoring of the pandemic. By processing large amounts of data and providing relevant information, ChatGPT has the potential to provide accurate and timely insights, and support decision-making processes to mitigate the spread and impact of pandemics. The paper highlights the importance of exploring the capabilities of artificial intelligence in the management of public emergencies and provides a starting point for future research in this area.
Collapse
Affiliation(s)
- Alberto G Gerli
- Department of Clinical Sciences and Community Health, University of Milan, Milan, Italy -
| | - Joan B Soriano
- Unit of Pulmonology, Hospital Universitario de la Princesa, Madrid, Spain
- Faculty of Medicine, Autonomous University of Madrid, Madrid, Spain
- Centro de Investigación Biomédica en Red de Enfermedades Respiratorias (CIBERES), Carlos III Health Institute, Madrid, Spain
| | - Gianfranco Alicandro
- Department of Pathophysiology and Transplantation, University of Milan, Milan, Italy
- Cystic Fibrosis Center, Fondazione IRCCS Ca' Granda Ospedale Maggiore Policlinico, Milan, Italy
| | - Michele Salvagno
- Department of Intensive Care, Erasme Hospital, Université Libre de Bruxelles, Brussels, Belgium
| | - Fabio Taccone
- Department of Intensive Care, Erasme Hospital, Université Libre de Bruxelles, Brussels, Belgium
| | - Stefano Centanni
- Respiratory Unit, Department of Health Sciences, ASST Santi Paolo e Carlo, University of Milan, Milan, Italy
| | - Carlo LA Vecchia
- Department of Clinical Sciences and Community Health, University of Milan, Milan, Italy
| |
Collapse
|
126
|
Bisi T, Risser A, Clavert P, Migaud H, Dartus J. What is the rate of text generated by artificial intelligence over a year of publication in Orthopedics & Traumatology: Surgery & Research? Analysis of 425 articles before versus after the launch of ChatGPT in November 2022. Orthop Traumatol Surg Res 2023; 109:103694. [PMID: 37776949 DOI: 10.1016/j.otsr.2023.103694] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/17/2023] [Revised: 07/10/2023] [Accepted: 08/24/2023] [Indexed: 10/02/2023]
Abstract
BACKGROUND The use of artificial intelligence (AI) is soaring, and the launch of ChatGPT in November 2022 has accelerated this trend. This "chatbot" can generate complete scientific articles, with risk of plagiarism by mining existing data or downright fraud by fabricating studies with no real data at all. There are tools that detect AI in publications, but to our knowledge they have not been systematically assessed for publication in scientific journals. We therefore conducted a retrospective study on articles published in Orthopaedics & Traumatology: Surgery & Research (OTSR): firstly, to screen for AI-generated content before and after the publicized launch of ChatGPT; secondly, to assess whether AI was more often used in some countries than others to generate content; thirdly, to determine whether plagiarism rate correlated with AI-generation, and lastly, to determine whether elements other than text generation, and notably the translation procedure, could raise suspicion of AI use. HYPOTHESIS The rate of AI use increased after the publicized launch of ChatGPT v3.5 in November 2022. MATERIAL AND METHODS In all, 425 articles published between February 2022 and September 2023 (221 before and 204 after November 1, 2022) underwent ZeroGPT assessment of the level of AI generation in the final English-language version (abstract and body of the article). Two scores were obtained: probability of AI generation, in six grades from Human to AI; and percentage AI generation. Plagiarism was assessed on the Ithenticate application at submission. Articles in French were assessed in their English-language version as translated by a human translator, with comparison to automatic translation by Google Translate and DeepL. RESULTS AI-generated text was detected mainly in Abstracts, with a 10.1% rate of AI or considerable AI generation, compared to only 1.9% for the body of the article and 5.6% for the total body+abstract. Analysis for before and after November 2022 found an increase in AI generation in body+abstract, from 10.30±15.95% (range, 0-100%) to 15.64±19.8% (range, 0-99.93) (p < 0.04; NS for abstracts alone). AI scores differed between types of article: 14.9% for original articles and 9.8% for reviews (p<0.01). The highest rates of probable AI generation were in articles from Japan, China, South America and English-speaking countries (p<0.0001). Plagiarism rates did not increase between the two study periods, and were unrelated to AI rates. On the other hand, when articles were classified as "suspected" of AI generation (plagiarism rate ≥ 20%) or "non-suspected" (rate<20%), the "similarity" score was higher in suspect articles: 25.7±13.23% (range, 10-69%) versus 16.28±10% (range, 0-79%) (p < 0.001). In the body of the article, use of translation software was associated with higher AI rates than with a human translator: 3.5±5% for human translators, versus 18±10% and 21.9±11% respectively for Google Translate and DeepL (p < 0.001). DISCUSSION The present study revealed an increasing rate of AI use in articles published in OTSR. AI grades differed according to type of article and country of origin. Use of translation software increased the AI grade. In the long run, use of ChatGPT incurs a risk of plagiarism and scientific misconduct, and needs to be detected and signaled by a digital tag on any robot-generated text. LEVEL OF EVIDENCE III; case-control study.
Collapse
Affiliation(s)
- Théophile Bisi
- Département universitaire de chirurgie orthopédique, université de Lille, CHU de Lille, 59000 Lille, France; Service de chirurgie orthopédique, centre hospitalier universitaire (CHU) de Lille, hôpital Roger-Salengro, place de Verdun, 59000 Lille, France.
| | - Anthony Risser
- Service de chirurgie du membre supérieur, Hautepierre 2, CHRU Strasbourg, 1, avenue Molière, 67200 Strasbourg, France
| | - Philippe Clavert
- Service de chirurgie du membre supérieur, Hautepierre 2, CHRU Strasbourg, 1, avenue Molière, 67200 Strasbourg, France; Faculté de médecine, institut d'anatomie normale, 4, rue Kirschleger, 67085 Strasbourg, France
| | - Henri Migaud
- Département universitaire de chirurgie orthopédique, université de Lille, CHU de Lille, 59000 Lille, France; Service de chirurgie orthopédique, centre hospitalier universitaire (CHU) de Lille, hôpital Roger-Salengro, place de Verdun, 59000 Lille, France
| | - Julien Dartus
- Département universitaire de chirurgie orthopédique, université de Lille, CHU de Lille, 59000 Lille, France; Service de chirurgie orthopédique, centre hospitalier universitaire (CHU) de Lille, hôpital Roger-Salengro, place de Verdun, 59000 Lille, France
| |
Collapse
|
127
|
Sabet CJ, Bajaj SS, Stanford FC, Celi LA. Equity in Scientific Publishing: Can Artificial Intelligence Transform the Peer Review Process? MAYO CLINIC PROCEEDINGS. DIGITAL HEALTH 2023; 1:596-600. [PMID: 40206303 PMCID: PMC11975676 DOI: 10.1016/j.mcpdig.2023.10.002] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 04/11/2025]
Affiliation(s)
- Cameron John Sabet
- Department of Medicine, Georgetown University School of Medicine, Washington, DC
| | - Simar S. Bajaj
- Department of Medicine, Harvard University, Cambridge, MA
| | - Fatima Cody Stanford
- Massachusetts General Hospital, MGH Weight Center, Department of Medicine-Division of Endocrinology-Neuroendocrine, Department of Pediatrics-Division of Endocrinology, Nutrition Obesity Research Center at Harvard (NORCH), Harvard Medical School, Boston, MA
| | - Leo Anthony Celi
- Department of Medicine, Beth Israel Deaconess Medical Center, Boston, MA
| |
Collapse
|
128
|
Mondal H, Mondal S. ChatGPT in academic writing: Maximizing its benefits and minimizing the risks. Indian J Ophthalmol 2023; 71:3600-3606. [PMID: 37991290 PMCID: PMC10788737 DOI: 10.4103/ijo.ijo_718_23] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2023] [Revised: 08/11/2023] [Accepted: 08/21/2023] [Indexed: 11/23/2023] Open
Abstract
This review article explores the use of ChatGPT in academic writing and provides insights on how to utilize it judiciously. With the increasing popularity of AI-powered language models, ChatGPT has emerged as a potential tool for assisting writers in the research and writing process. We have provided a list of potential uses of ChatGPT by a novice researcher for getting help during research proposal preparation and manuscript writing. However, there are concerns regarding its reliability and potential risks associated with its use. The review highlights the importance of maintaining human judgment in the writing process and using ChatGPT as a complementary tool rather than a replacement for human effort. The article concludes with recommendations for researchers and writers to ensure responsible and effective use of ChatGPT in academic writing.
Collapse
Affiliation(s)
- Himel Mondal
- Department of Physiology, All India Institute of Medical Sciences, Deoghar, Jharkhand, India
| | - Shaikat Mondal
- Department of Physiology, Raiganj Government Medical College and Hospital, West Bengal, India
| |
Collapse
|
129
|
Maroteau G, An JS, Murgier J, Hulet C, Ollivier M, Ferreira A. Evaluation of the impact of large language learning models on articles submitted to Orthopaedics & Traumatology: Surgery & Research (OTSR): A significant increase in the use of artificial intelligence in 2023. Orthop Traumatol Surg Res 2023; 109:103720. [PMID: 37866509 DOI: 10.1016/j.otsr.2023.103720] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/04/2023] [Revised: 10/04/2023] [Accepted: 10/04/2023] [Indexed: 10/24/2023]
Abstract
INTRODUCTION There has been an unprecedented rise is the use of artificial intelligence (AI) amongst medical fields. Recently, a dialogue agent called ChatGPT (Generative Pre-trained Transformer) has grown in popularity through its use of large language models (LLM) to clearly and precisely generate text on demand. However, the impact of AI on the creation of scientific articles is remains unknown. A retrospective study was carried out with the aim of answering the following questions: identify the presence of text generated by LLM before and after the increased usage of ChatGPT in articles submitted in OTSR; determine if the type of article, the year of submission, and the country of origin, influenced the proportion of text generated, at least in part by AI. MATERIAL AND METHODS A total of 390 English articles were submitted to OTSR in January, February and March 2022 (n=204) and over the same months of 2023 (n=186) were analyzed. All articles were analyzed using the ZeroGPT tool, which provides an assumed rate of AI use expressed as a percentage. A comparison of the average rate of AI use was carried out between the articles submitted in 2022 and 2023. This comparison was repeated keeping only the articles with the highest percentage of suspected AI use (greater than 10 and 20%). A secondary analysis was carried out to identify risk factors for AI use. RESULTS The average percentage of suspected LLM use in the entire cohort was 11%±6, with 160 articles (41.0%) having a suspected AI rate greater than 10% and 61 (15.6%) with an assumed AI rate greater than 20%. A comparison between articles submitted in 2022 and 2023 revealed a significant increase in the use of these tools after the launch of ChatGPT 3.5 (9.4% in 2022 and 12.6% in 2023 [p=0.004]). The number of articles with suspected AI rates of greater than 10 and 20% were significantly higher in 2023: >10%: 71 articles (34.8%) versus 89 articles (47.8%) (p=0.008) and >20%: 21 articles (10.3%) versus 40 articles (21.5%) (p=0.002). A risk factor analysis for LLLM use, demonstrated that authors of Asian geographic origin, and the submission year 2023 were associated with a higher rate of suspected AI use. An AI rate >20% was associated to Asian geographical origin with an odds ratio of 1.79 (95% CI: 1.03-3.11) (p=0.029), while the year of submission being 2023 had an odds ratio of 1.7 (95% CI: 1.1-2.5) (p=0.02). CONCLUSION This study highlights a significant increase in the use of LLM in the writing of articles submitted to the OTSR journal after the launch of ChatGPT 3.5. The increasing use of these models raises questions about originality and plagiarism in scientific research. AI offers creative opportunities but also raises ethical and methodological challenges. LEVEL OF EVIDENCE III; case control study.
Collapse
Affiliation(s)
- Gaëlle Maroteau
- Unité Inserm Comète 1075, Department of Orthopaedics and Traumatology, Caen University Hospital, avenue Cote-de-Nacre, 14000 Caen, France
| | - Jae-Sung An
- Tokyo Medical and Dental University, 1 Chome-5-45 Yushima, Bunkyo City, Tokyo 113-8510, Japan
| | - Jérome Murgier
- Service de chirurgie orthopédique, clinique Aguiléra, 21, rue de l'Estagnas, 64200 Biarritz, France
| | - Christophe Hulet
- Unité Inserm Comète 1075, Department of Orthopaedics and Traumatology, Caen University Hospital, avenue Cote-de-Nacre, 14000 Caen, France
| | - Matthieu Ollivier
- Institute of Movement and Locomotion, Department of Orthopaedics and Traumatology, Sainte-Marguerite Hospital, BP 29, 270, boulevard Sainte-Marguerite, 13274 Marseille, France; Aix-Marseille Unit, Institute for Locomotion, Department of Orthopaedics and Traumatology, CNRS, ISM, Sainte-Marguerite Hospital, AP-HM, Marseille, France
| | - Alexandre Ferreira
- Unité Inserm Comète 1075, Department of Orthopaedics and Traumatology, Caen University Hospital, avenue Cote-de-Nacre, 14000 Caen, France.
| |
Collapse
|
130
|
Jacques T, Sleiman R, Diaz MI, Dartus J. Artificial intelligence: Emergence and possible fraudulent use in medical publishing. Orthop Traumatol Surg Res 2023; 109:103709. [PMID: 37852535 DOI: 10.1016/j.otsr.2023.103709] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/25/2023] [Accepted: 10/06/2023] [Indexed: 10/20/2023]
Affiliation(s)
- Thibaut Jacques
- IRIS Imagerie, 144, avenue de Dunkerque, 59000 Lille, France.
| | - Rita Sleiman
- Centre de recherche et d'innovation de Talan, 14, rue Pergolèse, 75116 Paris, France
| | - Manuel I Diaz
- Centre de recherche et d'innovation de Talan, 14, rue Pergolèse, 75116 Paris, France
| | - Julien Dartus
- Département universitaire de chirurgie orthopédique et traumatologique, hôpital Roger-Salengro, CHU de Lille ULR 4490, université de Lille, place de Verdun, 59037 Lille, France; U1008 - Controlled Drug Delivery Systems and Biomaterials, CHU de Lille, University Lille, Inserm, 59000 Lille, France
| |
Collapse
|
131
|
Kantor J. ChatGPT, large language models, and artificial intelligence in medicine and health care: A primer for clinicians and researchers. JAAD Int 2023; 13:168-169. [PMID: 37823044 PMCID: PMC10562174 DOI: 10.1016/j.jdin.2023.07.011] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/13/2023] Open
Affiliation(s)
- Jonathan Kantor
- Department of Dermatology, Center for Global Health, and Center for Clinical Epidemiology and Biostatistics, Perelman School of Medicine, University of Pennsylvania, Philadelphia, Pennsylvania; Florida Center for Dermatology, St Augustine, Florida; and Alchemy Labs, Oxford, UK
| |
Collapse
|
132
|
Hryciw BN, Seely AJE, Kyeremanteng K. Guiding principles and proposed classification system for the responsible adoption of artificial intelligence in scientific writing in medicine. Front Artif Intell 2023; 6:1283353. [PMID: 38035200 PMCID: PMC10687472 DOI: 10.3389/frai.2023.1283353] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2023] [Accepted: 10/17/2023] [Indexed: 12/02/2023] Open
Abstract
The integration of large language models (LLMs) and artificial intelligence (AI) into scientific writing, especially in medical literature, presents both unprecedented opportunities and inherent challenges. This manuscript evaluates the transformative potential of LLMs for the synthesis of information, linguistic enhancements, and global knowledge dissemination. At the same time, it raises concerns about unintentional plagiarism, the risk of misinformation, data biases, and an over-reliance on AI. To address these, we propose governing principles for AI adoption that ensure integrity, transparency, validity, and accountability. Additionally, guidelines for reporting AI involvement in manuscript development are delineated, and a classification system to specify the level of AI assistance is introduced. This approach uniquely addresses the challenges of AI in scientific writing, emphasizing transparency in authorship, qualification of AI involvement, and ethical considerations. Concerns regarding access equity, potential biases in AI-generated content, authorship dynamics, and accountability are also explored, emphasizing the human author's continued responsibility. Recommendations are made for fostering collaboration between AI developers, researchers, and journal editors and for emphasizing the importance of AI's responsible use in academic writing. Regular evaluations of AI's impact on the quality and biases of medical manuscripts are also advocated. As we navigate the expanding realm of AI in scientific discourse, it is crucial to maintain the human element of creativity, ethics, and oversight, ensuring that the integrity of scientific literature remains uncompromised.
Collapse
Affiliation(s)
- Brett N. Hryciw
- Division of Critical Care, Department of Medicine, University of Ottawa, Ottawa, ON, Canada
| | - Andrew J. E. Seely
- Division of Thoracic Surgery, Department of Surgery, The Ottawa Hospital, Ottawa, ON, Canada
- Clinical Epidemiology, Ottawa Hospital Research Institute, University of Ottawa, Ottawa, ON, Canada
| | - Kwadwo Kyeremanteng
- Division of Critical Care, Department of Medicine, University of Ottawa, Ottawa, ON, Canada
- Clinical Epidemiology, Ottawa Hospital Research Institute, University of Ottawa, Ottawa, ON, Canada
- Institute du Savoir Montfort, Ottawa, ON, Canada
| |
Collapse
|
133
|
Gödde D, Nöhl S, Wolf C, Rupert Y, Rimkus L, Ehlers J, Breuckmann F, Sellmann T. A SWOT (Strengths, Weaknesses, Opportunities, and Threats) Analysis of ChatGPT in the Medical Literature: Concise Review. J Med Internet Res 2023; 25:e49368. [PMID: 37865883 PMCID: PMC10690535 DOI: 10.2196/49368] [Citation(s) in RCA: 15] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2023] [Revised: 09/26/2023] [Accepted: 09/27/2023] [Indexed: 10/23/2023] Open
Abstract
BACKGROUND ChatGPT is a 175-billion-parameter natural language processing model that is already involved in scientific content and publications. Its influence ranges from providing quick access to information on medical topics, assisting in generating medical and scientific articles and papers, performing medical data analyses, and even interpreting complex data sets. OBJECTIVE The future role of ChatGPT remains uncertain and a matter of debate already shortly after its release. This review aimed to analyze the role of ChatGPT in the medical literature during the first 3 months after its release. METHODS We performed a concise review of literature published in PubMed from December 1, 2022, to March 31, 2023. To find all publications related to ChatGPT or considering ChatGPT, the search term was kept simple ("ChatGPT" in AllFields). All publications available as full text in German or English were included. All accessible publications were evaluated according to specifications by the author team (eg, impact factor, publication modus, article type, publication speed, and type of ChatGPT integration or content). The conclusions of the articles were used for later SWOT (strengths, weaknesses, opportunities, and threats) analysis. All data were analyzed on a descriptive basis. RESULTS Of 178 studies in total, 160 met the inclusion criteria and were evaluated. The average impact factor was 4.423 (range 0-96.216), and the average publication speed was 16 (range 0-83) days. Among the articles, there were 77 editorials (48,1%), 43 essays (26.9%), 21 studies (13.1%), 6 reviews (3.8%), 6 case reports (3.8%), 6 news (3.8%), and 1 meta-analysis (0.6%). Of those, 54.4% (n=87) were published as open access, with 5% (n=8) provided on preprint servers. Over 400 quotes with information on strengths, weaknesses, opportunities, and threats were detected. By far, most (n=142, 34.8%) were related to weaknesses. ChatGPT excels in its ability to express ideas clearly and formulate general contexts comprehensibly. It performs so well that even experts in the field have difficulty identifying abstracts generated by ChatGPT. However, the time-limited scope and the need for corrections by experts were mentioned as weaknesses and threats of ChatGPT. Opportunities include assistance in formulating medical issues for nonnative English speakers, as well as the possibility of timely participation in the development of such artificial intelligence tools since it is in its early stages and can therefore still be influenced. CONCLUSIONS Artificial intelligence tools such as ChatGPT are already part of the medical publishing landscape. Despite their apparent opportunities, policies and guidelines must be implemented to ensure benefits in education, clinical practice, and research and protect against threats such as scientific misconduct, plagiarism, and inaccuracy.
Collapse
Affiliation(s)
- Daniel Gödde
- Department of Pathology and Molecularpathology, Helios University Hospital Wuppertal, Witten/Herdecke University, Witten, Germany
| | - Sophia Nöhl
- Faculty of Health, Witten/Herdecke University, Witten, Germany
| | - Carina Wolf
- Faculty of Health, Witten/Herdecke University, Witten, Germany
| | - Yannick Rupert
- Faculty of Health, Witten/Herdecke University, Witten, Germany
| | - Lukas Rimkus
- Faculty of Health, Witten/Herdecke University, Witten, Germany
| | - Jan Ehlers
- Department of Didactics and Education Research in the Health Sector, Faculty of Health, Witten/Herdecke University, Witten, Germany
| | - Frank Breuckmann
- Department of Cardiology and Vascular Medicine, West German Heart and Vascular Center Essen, University Duisburg-Essen, Essen, Germany
- Department of Cardiology, Pneumology, Neurology and Intensive Care Medicine, Klinik Kitzinger Land, Kitzingen, Germany
| | - Timur Sellmann
- Department of Anaesthesiology I, Witten/Herdecke University, Witten, Germany
- Department of Anaesthesiology and Intensive Care Medicine, Evangelisches Krankenhaus BETHESDA zu Duisburg, Duisburg, Germany
| |
Collapse
|
134
|
Wang X, Liu XQ. Potential and limitations of ChatGPT and generative artificial intelligence in medical safety education. World J Clin Cases 2023; 11:7935-7939. [DOI: 10.12998/wjcc.v11.i32.7935] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/07/2023] [Revised: 09/21/2023] [Accepted: 11/02/2023] [Indexed: 11/16/2023] Open
Abstract
The primary objectives of medical safety education are to provide the public with essential knowledge about medications and to foster a scientific approach to drug usage. The era of using artificial intelligence to revolutionize medical safety education has already dawned, and ChatGPT and other generative artificial intelligence models have immense potential in this domain. Notably, they offer a wealth of knowledge, anonymity, continuous availability, and personalized services. However, the practical implementation of generative artificial intelligence models such as ChatGPT in medical safety education still faces several challenges, including concerns about the accuracy of information, legal responsibilities, and ethical obligations. Moving forward, it is crucial to intelligently upgrade ChatGPT by leveraging the strengths of existing medical practices. This task involves further integrating the model with real-life scenarios and proactively addressing ethical and security issues with the ultimate goal of providing the public with comprehensive, convenient, efficient, and personalized medical services.
Collapse
Affiliation(s)
- Xin Wang
- School of Education, Tianjin University, Tianjin 300350, China
| | - Xin-Qiao Liu
- School of Education, Tianjin University, Tianjin 300350, China
| |
Collapse
|
135
|
He S, Yang F, Zuo JP, Lin ZM. ChatGPT for scientific paper writing-promises and perils. Innovation (N Y) 2023; 4:100524. [PMID: 38028132 PMCID: PMC10654578 DOI: 10.1016/j.xinn.2023.100524] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2023] [Accepted: 10/17/2023] [Indexed: 12/01/2023] Open
Affiliation(s)
- Shijun He
- Innovation Research Institute of Traditional Chinese Medicine, Shanghai University of Traditional Chinese Medicine, Shanghai 201203, China
| | - Fan Yang
- Shanghai Key Laboratory of Gynecologic Oncology, Ren Ji Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai 200127, China
| | - Jian-ping Zuo
- Laboratory of Immunopharmacology, State Key Laboratory of Drug Research, Shanghai Institute of Materia Medica, Chinese Academy of Sciences, Shanghai 201203, China
| | - Ze-min Lin
- Laboratory of Immunopharmacology, State Key Laboratory of Drug Research, Shanghai Institute of Materia Medica, Chinese Academy of Sciences, Shanghai 201203, China
| |
Collapse
|
136
|
Rivera H. [Scientific integrity faces plagiarism fabricated with the ChatGPT]. REVISTA MEDICA DEL INSTITUTO MEXICANO DEL SEGURO SOCIAL 2023; 61:857-862. [PMID: 37995379 PMCID: PMC10723832] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Subscribe] [Scholar Register] [Received: 04/10/2023] [Accepted: 06/13/2023] [Indexed: 11/25/2023]
Abstract
Among the malpractices that undermine research integrity, plagiarism is a major threat given its frequency and evolving presentations. Plagiarism implies the intentional grabbing of texts, ideas, images, or data belonging to others and without crediting them. However, the different and even masked forms of plagiarism often difficult a clear identification. Currently, the many kinds of fraud and plagiarism account for most retractions in traditional and open access journals. Further, the rate of retracted articles is higher in the Latin American databases LILACS and Scielo than in PubMed and Web of Science. This difference has been related to the typical laxity of our culture and the lack of English writing skills of non-Anglophone researchers. These features explain the conflict experienced by Latin American students in USA where they face a stricter culture regarding academic and scientific plagiarism. In the internet era, the ease of accessing scientific literature has increased the temptation to plagiarize but this ethical breach has been countered by antiplagiarism software. Now, the so-called "paraphragiarism" prompted by paraphrasing tools exceeds the infamous "copy-paste". For instance, the innovative ChatGPT can be used for plagiarizing and paraphragiarizing. Moreover, its inclusion as coauthor in scientific papers has been banned by prestigious journals and the International Committee of Medical Journal Editors because such chatbot cannot meet the required public responsibility criterium. To avoid plagiarism, it is enough to always give due credit in the proper way. Lastly, I question the ill-fated and now prevailing conjunction of blind faith in progress and zero skepticism that prevents us from foreseeing the negative consequences of technological advances.
Collapse
Affiliation(s)
- Horacio Rivera
- Universidad de Guadalajara, Centro Universitario de Ciencias de la Salud, Departamento de Biología Molecular y Genómica. Guadalajara, Jalisco, MéxicoUniversidad de GuadalajaraMéxico
| |
Collapse
|
137
|
Chlorogiannis DD, Apostolos A, Chlorogiannis A, Palaiodimos L, Giannakoulas G, Pargaonkar S, Xesfingi S, Kokkinidis DG. The Role of ChatGPT in the Advancement of Diagnosis, Management, and Prognosis of Cardiovascular and Cerebrovascular Disease. Healthcare (Basel) 2023; 11:2906. [PMID: 37958050 PMCID: PMC10648908 DOI: 10.3390/healthcare11212906] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/03/2023] [Revised: 10/24/2023] [Accepted: 11/04/2023] [Indexed: 11/15/2023] Open
Abstract
Cardiovascular and cerebrovascular disease incidence has risen mainly due to poor control of preventable risk factors and still constitutes a significant financial and health burden worldwide. ChatGPT is an artificial intelligence language-based model developed by OpenAI. Due to the model's unique cognitive capabilities beyond data processing and the production of high-quality text, there has been a surge of research interest concerning its role in the scientific community and contemporary clinical practice. To fully exploit ChatGPT's potential benefits and reduce its possible misuse, extreme caution must be taken to ensure its implications ethically and equitably. In this narrative review, we explore the language model's possible applications and limitations while emphasizing its potential value for diagnosing, managing, and prognosis of cardiovascular and cerebrovascular disease.
Collapse
Affiliation(s)
| | - Anastasios Apostolos
- First Department of Cardiology, School of Medicine, National Kapodistrian University of Athens, Hippokrateion General Hospital of Athens, 115 27 Athens, Greece;
| | - Anargyros Chlorogiannis
- Department of Health Economics, Policy and Management, Karolinska Institutet, 171 77 Stockholm, Sweden
| | - Leonidas Palaiodimos
- Division of Hospital Medicine, Jacobi Medical Center, NYC H+H, Albert Einstein College of Medicine, New York, NY 10461, USA; (L.P.); (S.P.)
| | - George Giannakoulas
- Department of Cardiology, AHEPA University Hospital, Aristotle University of Thessaloniki, 541 24 Thessaloniki, Greece;
| | - Sumant Pargaonkar
- Division of Hospital Medicine, Jacobi Medical Center, NYC H+H, Albert Einstein College of Medicine, New York, NY 10461, USA; (L.P.); (S.P.)
| | - Sofia Xesfingi
- Department of Economics, University of Piraeus, 185 34 Piraeus, Greece
| | - Damianos G. Kokkinidis
- Section of Cardiovascular Medicine, Yale University School of Medicine, New Haven, CT 06510, USA
| |
Collapse
|
138
|
Botelho F, Tshimula JM, Poenaru D. Leveraging ChatGPT to Democratize and Decolonize Global Surgery: Large Language Models for Small Healthcare Budgets. World J Surg 2023; 47:2626-2627. [PMID: 37689598 DOI: 10.1007/s00268-023-07167-2] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 08/18/2023] [Indexed: 09/11/2023]
Affiliation(s)
- Fabio Botelho
- Departments of Pediatric Surgery and Surgical and Interventional Sciences, McGill University, Montreal, Canada
| | - Jean Marie Tshimula
- Departments of Pediatric Surgery and Surgical and Interventional Sciences, McGill University, Montreal, Canada
| | - Dan Poenaru
- Departments of Pediatric Surgery and Surgical and Interventional Sciences, McGill University, Montreal, Canada.
| |
Collapse
|
139
|
Chandra A, Dasgupta S. Impact of ChatGPT on Medical Research Article Writing and Publication. Sultan Qaboos Univ Med J 2023; 23:429-432. [PMID: 38090250 PMCID: PMC10712376 DOI: 10.18295/squmj.11.2023.068] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2023] [Revised: 10/14/2023] [Accepted: 10/15/2023] [Indexed: 12/18/2023] Open
Affiliation(s)
- Atanu Chandra
- Department of Internal Medicine, Bankura Sammilani Medical College and Hospital, West Bengal, India
| | - Sugata Dasgupta
- Department of Critical Care Medicine, IPGMER and SSKM Hospital, West Bengal, India
| |
Collapse
|
140
|
Sikander B, Baker JJ, Deveci CD, Lund L, Rosenberg J. ChatGPT-4 and Human Researchers Are Equal in Writing Scientific Introduction Sections: A Blinded, Randomized, Non-inferiority Controlled Study. Cureus 2023; 15:e49019. [PMID: 38111405 PMCID: PMC10727453 DOI: 10.7759/cureus.49019] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/18/2023] [Indexed: 12/20/2023] Open
Abstract
Background Natural language processing models are increasingly used in scientific research, and their ability to perform various tasks in the research process is rapidly advancing. This study aims to investigate whether Generative Pre-trained Transformer 4 (GPT-4) is equal to humans in writing introduction sections for scientific articles. Methods This randomized non-inferiority study was reported according to the Consolidated Standards of Reporting Trials for non-inferiority trials and artificial intelligence (AI) guidelines. GPT-4 was instructed to synthesize 18 introduction sections based on the aim of previously published studies, and these sections were compared to the human-written introductions already published in a medical journal. Eight blinded assessors randomly evaluated the introduction sections using 1-10 Likert scales. Results There was no significant difference between GPT-4 and human introductions regarding publishability and content quality. GPT-4 had one point significantly better scores in readability, which was considered a non-relevant difference. The majority of assessors (59%) preferred GPT-4, while 33% preferred human-written introductions. Based on Lix and Flesch-Kincaid scores, GPT-4 introductions were 10 and two points higher, respectively, indicating that the sentences were longer and had longer words. Conclusion GPT-4 was found to be equal to humans in writing introductions regarding publishability, readability, and content quality. The majority of assessors preferred GPT-4 introductions and less than half could determine which were written by GPT-4 or humans. These findings suggest that GPT-4 can be a useful tool for writing introduction sections, and further studies should evaluate its ability to write other parts of scientific articles.
Collapse
Affiliation(s)
| | | | | | - Lars Lund
- Urology, Odense University Hospital, Odense, DNK
| | | |
Collapse
|
141
|
Yan M, Cerri GG, Moraes FY. ChatGPT and medicine: how AI language models are shaping the future and health related careers. Nat Biotechnol 2023; 41:1657-1658. [PMID: 37950005 DOI: 10.1038/s41587-023-02011-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2023]
Affiliation(s)
- Michael Yan
- Radiation Medicine Program, Princess Margaret Cancer Centre-University Health Network, Toronto, Ontario, Canada
| | - Giovanni G Cerri
- Department of Radiology and Oncology, University of São Paulo Medical School, São Paulo, Brazil
| | - Fabio Y Moraes
- Department of Radiology and Oncology, University of São Paulo Medical School, São Paulo, Brazil.
- Cancer Care and Epidemiology, Cancer Research Institute, Queen's University, Kingston, Ontario, Canada.
- Division of Radiation Oncology, Kingston General Hospital, Queen's University, Kingston, Ontario, Canada.
| |
Collapse
|
142
|
Arillotta D, Floresta G, Guirguis A, Corkery JM, Catalani V, Martinotti G, Sensi SL, Schifano F. GLP-1 Receptor Agonists and Related Mental Health Issues; Insights from a Range of Social Media Platforms Using a Mixed-Methods Approach. Brain Sci 2023; 13:1503. [PMID: 38002464 PMCID: PMC10669484 DOI: 10.3390/brainsci13111503] [Citation(s) in RCA: 23] [Impact Index Per Article: 11.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2023] [Revised: 10/16/2023] [Accepted: 10/23/2023] [Indexed: 11/26/2023] Open
Abstract
The emergence of glucagon-like peptide-1 receptor agonists (GLP-1 RAs; semaglutide and others) now promises effective, non-invasive treatment of obesity for individuals with and without diabetes. Social media platforms' users started promoting semaglutide/Ozempic as a weight-loss treatment, and the associated increase in demand has contributed to an ongoing worldwide shortage of the drug associated with levels of non-prescribed semaglutide intake. Furthermore, recent reports emphasized some GLP-1 RA-associated risks of triggering depression and suicidal thoughts. Consistent with the above, we aimed to assess the possible impact of GLP-1 RAs on mental health as being perceived and discussed in popular open platforms with the help of a mixed-methods approach. Reddit posts yielded 12,136 comments, YouTube videos 14,515, and TikTok videos 17,059, respectively. Out of these posts/entries, most represented matches related to sleep-related issues, including insomnia (n = 620 matches); anxiety (n = 353); depression (n = 204); and mental health issues in general (n = 165). After the initiation of GLP-1 RAs, losing weight was associated with either a marked improvement or, in some cases, a deterioration, in mood; increase/decrease in anxiety/insomnia; and better control of a range of addictive behaviors. The challenges of accessing these medications were a hot topic as well. To the best of our knowledge, this is the first study documenting if and how GLP-1 RAs are perceived as affecting mood, mental health, and behaviors. Establishing a clear cause-and-effect link between metabolic diseases, depression and medications is difficult because of their possible reciprocal relationship, shared underlying mechanisms and individual differences. Further research is needed to better understand the safety profile of these molecules and their putative impact on behavioral and non-behavioral addictions.
Collapse
Affiliation(s)
- Davide Arillotta
- School of Clinical Pharmacology and Toxicology, University of Florence, 50121 Florence, Italy;
- Psychopharmacology, Drug Misuse and Novel Psychoactive Substances Research Unit, School of Life and Medical Sciences, University of Hertfordshire, Hatfield AL10 9AB, UK; (G.F.); (A.G.); (J.M.C.); (V.C.); (G.M.)
| | - Giuseppe Floresta
- Psychopharmacology, Drug Misuse and Novel Psychoactive Substances Research Unit, School of Life and Medical Sciences, University of Hertfordshire, Hatfield AL10 9AB, UK; (G.F.); (A.G.); (J.M.C.); (V.C.); (G.M.)
- Department of Drug and Health Sciences, University of Catania, 95124 Catania, Italy
| | - Amira Guirguis
- Psychopharmacology, Drug Misuse and Novel Psychoactive Substances Research Unit, School of Life and Medical Sciences, University of Hertfordshire, Hatfield AL10 9AB, UK; (G.F.); (A.G.); (J.M.C.); (V.C.); (G.M.)
- Pharmacy, Swansea University Medical School, Faculty of Medicine, Health and Life Science, Swansea University, Swansea SA2 8PP, UK
| | - John Martin Corkery
- Psychopharmacology, Drug Misuse and Novel Psychoactive Substances Research Unit, School of Life and Medical Sciences, University of Hertfordshire, Hatfield AL10 9AB, UK; (G.F.); (A.G.); (J.M.C.); (V.C.); (G.M.)
| | - Valeria Catalani
- Psychopharmacology, Drug Misuse and Novel Psychoactive Substances Research Unit, School of Life and Medical Sciences, University of Hertfordshire, Hatfield AL10 9AB, UK; (G.F.); (A.G.); (J.M.C.); (V.C.); (G.M.)
| | - Giovanni Martinotti
- Psychopharmacology, Drug Misuse and Novel Psychoactive Substances Research Unit, School of Life and Medical Sciences, University of Hertfordshire, Hatfield AL10 9AB, UK; (G.F.); (A.G.); (J.M.C.); (V.C.); (G.M.)
- Department of Neurosciences, Imaging and Clinical Sciences, University of Chieti-Pescara, 66100 Chieti, Italy;
| | - Stefano L. Sensi
- Department of Neurosciences, Imaging and Clinical Sciences, University of Chieti-Pescara, 66100 Chieti, Italy;
- Center for Advanced Studies and Technology (CAST), Institute of Advanced Biomedical Technology (ITAB), University of Chieti-Pescara, Via dei Vestini 21, 66100 Chieti, Italy
| | - Fabrizio Schifano
- Psychopharmacology, Drug Misuse and Novel Psychoactive Substances Research Unit, School of Life and Medical Sciences, University of Hertfordshire, Hatfield AL10 9AB, UK; (G.F.); (A.G.); (J.M.C.); (V.C.); (G.M.)
| |
Collapse
|
143
|
Sridi C, Brigui S. The use of ChatGPT in occupational medicine: opportunities and threats. Ann Occup Environ Med 2023; 35:e42. [PMID: 38029273 PMCID: PMC10654530 DOI: 10.35371/aoem.2023.35.e42] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2023] [Revised: 09/30/2023] [Accepted: 10/06/2023] [Indexed: 12/01/2023] Open
Abstract
ChatGPT has the potential to revolutionize occupational medicine by providing a powerful tool for analyzing data, improving communication, and increasing efficiency. It can help identify patterns and trends in workplace health and safety, act as a virtual assistant for workers, employers, and occupational health professionals, and automate certain tasks. However, caution is required due to ethical concerns, the need to maintain confidentiality, and the risk of inconsistent or inaccurate results. ChatGPT cannot replace the crucial role of the occupational health professional in the medical surveillance of workers and the analysis of data on workers' health.
Collapse
Affiliation(s)
- Chayma Sridi
- Department of Occupational Medicine, Sahloul University Hospital, Sousse, Tunisia
| | | |
Collapse
|
144
|
Liu G, Ma X, Zhang Y, Su B, Liu P. GPT4: The Indispensable Helper for Neurosurgeons in the New Era. Ann Biomed Eng 2023; 51:2113-2115. [PMID: 37204548 DOI: 10.1007/s10439-023-03241-x] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2023] [Accepted: 05/11/2023] [Indexed: 05/20/2023]
Abstract
GPT4 is the newest multimodal language model released by OpenAI. With its powerful capabilities, GPT4 has great potential to revolutionize the healthcare industry. In this study, we proposed various ways GPT4 could display its talents in the field of neurosurgery in future. We believe that GPT4 is prone to become an indispensable assistant for neurosurgeons in the new era.
Collapse
Affiliation(s)
- Gemingtian Liu
- Department of Neurosurgery, Beijing Tiantan Hospital, Capital Medical University, Beijing, China
| | - Xin Ma
- Department of Neurosurgery, Beijing Tiantan Hospital, Capital Medical University, Beijing, China
| | - Yu Zhang
- Department of Neurosurgery, Beijing Tiantan Hospital, Capital Medical University, Beijing, China
| | - Boyan Su
- Department of Neurosurgery, Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Pinan Liu
- Department of Neurosurgery, Beijing Tiantan Hospital, Capital Medical University, Beijing, China.
- Department of Neural Reconstruction, Beijing Neurosurgery Institute, Capital Medical University, Beijing, China.
| |
Collapse
|
145
|
Diaz Milian R, Moreno Franco P, Freeman WD, Halamka JD. Revolution or Peril? The Controversial Role of Large Language Models in Medical Manuscript Writing. Mayo Clin Proc 2023; 98:1444-1448. [PMID: 37793723 DOI: 10.1016/j.mayocp.2023.07.009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/23/2023] [Accepted: 07/13/2023] [Indexed: 10/06/2023]
Affiliation(s)
- Ricardo Diaz Milian
- Division of Critical Care Medicine, Mayo Clinic, Jacksonville, FL; Department of Anesthesiology, Mayo Clinic, Jacksonville, FL.
| | - Pablo Moreno Franco
- Department of Anesthesiology, Mayo Clinic, Jacksonville, FL; Department of Transplant Medicine, Mayo Clinic, Jacksonville, FL
| | - William D Freeman
- Department of Neurology and Neurosurgery, Mayo Clinic, Jacksonville, FL
| | - John D Halamka
- Department of Emergency Medicine, Mayo Clinic, Rochester, MN; Department of Internal Medicine, Mayo Clinic, Rochester, MN; Mayo Clinic Platform, Mayo Clinic, Rochester, MN
| |
Collapse
|
146
|
Ergun Y. Redefining authorship in the era of artificial intelligence: balancing ethics, transparency, and progress. ESMO Open 2023; 8:101634. [PMID: 37659291 PMCID: PMC10480051 DOI: 10.1016/j.esmoop.2023.101634] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2023] [Accepted: 08/04/2023] [Indexed: 09/04/2023] Open
Affiliation(s)
- Y Ergun
- Department of Medical Oncology, Batman World Hospital, Batman, Turkey.
| |
Collapse
|
147
|
Huespe IA, Echeverri J, Khalid A, Carboni Bisso I, Musso CG, Surani S, Bansal V, Kashyap R. Clinical Research With Large Language Models Generated Writing-Clinical Research with AI-assisted Writing (CRAW) Study. Crit Care Explor 2023; 5:e0975. [PMID: 37795455 PMCID: PMC10547240 DOI: 10.1097/cce.0000000000000975] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/06/2023] Open
Abstract
IMPORTANCE The scientific community debates Generative Pre-trained Transformer (GPT)-3.5's article quality, authorship merit, originality, and ethical use in scientific writing. OBJECTIVES Assess GPT-3.5's ability to craft the background section of critical care clinical research questions compared to medical researchers with H-indices of 22 and 13. DESIGN Observational cross-sectional study. SETTING Researchers from 20 countries from six continents evaluated the backgrounds. PARTICIPANTS Researchers with a Scopus index greater than 1 were included. MAIN OUTCOMES AND MEASURES In this study, we generated a background section of a critical care clinical research question on "acute kidney injury in sepsis" using three different methods: researcher with H-index greater than 20, researcher with H-index greater than 10, and GPT-3.5. The three background sections were presented in a blinded survey to researchers with an H-index range between 1 and 96. First, the researchers evaluated the main components of the background using a 5-point Likert scale. Second, they were asked to identify which background was written by humans only or with large language model-generated tools. RESULTS A total of 80 researchers completed the survey. The median H-index was 3 (interquartile range, 1-7.25) and most (36%) researchers were from the Critical Care specialty. When compared with researchers with an H-index of 22 and 13, GPT-3.5 was marked high on the Likert scale ranking on main background components (median 4.5 vs. 3.82 vs. 3.6 vs. 4.5, respectively; p < 0.001). The sensitivity and specificity to detect researchers writing versus GPT-3.5 writing were poor, 22.4% and 57.6%, respectively. CONCLUSIONS AND RELEVANCE GPT-3.5 could create background research content indistinguishable from the writing of a medical researcher. It was marked higher compared with medical researchers with an H-index of 22 and 13 in writing the background section of a critical care clinical research question.
Collapse
Affiliation(s)
- Ivan A Huespe
- Hospital Italiano de Buenos Aires, Buenos Aires, Argentina
- Universidad de Buenos Aires, Buenos Aires, Argentina
| | | | | | | | - Carlos G Musso
- Hospital Italiano de Buenos Aires, Buenos Aires, Argentina
- Facultad de Ciencias de la Salud, Universidad Simon Bolivar, Barranquilla, Colombia
| | - Salim Surani
- Mayo Clinic, Rochester, MN
- Texas A&M University, College Station, TX
| | | | | |
Collapse
|
148
|
Cocci A, Pezzoli M, Minervini A. Light and Shadow of ChatGPT: A Real Tool for Advancing Scientific Research and Medical Practice? World J Mens Health 2023; 41:751-752. [PMID: 37652659 PMCID: PMC10523110 DOI: 10.5534/wjmh.230102] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2023] [Revised: 04/26/2023] [Accepted: 05/03/2023] [Indexed: 09/02/2023] Open
Affiliation(s)
- Andrea Cocci
- Department of Urology, Azienda Ospedaliera Universitaria Careggi, University of Florence, Florence, Italy
| | - Marta Pezzoli
- Department of Urology, Azienda Ospedaliera Universitaria Careggi, University of Florence, Florence, Italy.
| | - Andrea Minervini
- Department of Urology, Azienda Ospedaliera Universitaria Careggi, University of Florence, Florence, Italy
| |
Collapse
|
149
|
Tasker RC. Writing for Pediatric Critical Care Medicine: Engaging With Citations to References in the Chatbot Generative Pre-Trained Transformer Era. Pediatr Crit Care Med 2023; 24:862-868. [PMID: 38412368 DOI: 10.1097/pcc.0000000000003356] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 02/29/2024]
Affiliation(s)
- Robert C Tasker
- orcid.org/0000-0003-3647-8113
- Department of Anesthesiology, Critical Care and Pain Medicine, Boston Children's Hospital, Boston, MA
- Selwyn College, Cambridge University, Cambridge, United Kingdom
| |
Collapse
|
150
|
Hwang SI, Lim JS, Lee RW, Matsui Y, Iguchi T, Hiraki T, Ahn H. Is ChatGPT a "Fire of Prometheus" for Non-Native English-Speaking Researchers in Academic Writing? Korean J Radiol 2023; 24:952-959. [PMID: 37793668 PMCID: PMC10550740 DOI: 10.3348/kjr.2023.0773] [Citation(s) in RCA: 18] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2023] [Accepted: 08/18/2023] [Indexed: 10/06/2023] Open
Abstract
Large language models (LLMs) such as ChatGPT have garnered considerable interest for their potential to aid non-native English-speaking researchers. These models can function as personal, round-the-clock English tutors, akin to how Prometheus in Greek mythology bestowed fire upon humans for their advancement. LLMs can be particularly helpful for non-native researchers in writing the Introduction and Discussion sections of manuscripts, where they often encounter challenges. However, using LLMs to generate text for research manuscripts entails concerns such as hallucination, plagiarism, and privacy issues; to mitigate these risks, authors should verify the accuracy of generated content, employ text similarity detectors, and avoid inputting sensitive information into their prompts. Consequently, it may be more prudent to utilize LLMs for editing and refining text rather than generating large portions of text. Journal policies concerning the use of LLMs vary, but transparency in disclosing artificial intelligence tool usage is emphasized. This paper aims to summarize how LLMs can lower the barrier to academic writing in English, enabling researchers to concentrate on domain-specific research, provided they are used responsibly and cautiously.
Collapse
Affiliation(s)
- Sung Il Hwang
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
- Department of Radiology, Graduate School of Medicine, Dentistry and Pharmaceutical Sciences, Okayama University, Okayama, Japan
| | - Joon Seo Lim
- Scientific Publications Team, Clinical Research Center, Asan Medical Center, University of Ulsan College of Medicine, Seoul, Republic of Korea.
| | - Ro Woon Lee
- Department of Radiology, Inha University Hospital, Incheon, Republic of Korea
| | - Yusuke Matsui
- Department of Radiology, Faculty of Medicine, Dentistry and Pharmaceutical Sciences, Okayama University, Okayama, Japan
| | - Toshihiro Iguchi
- Department of Radiological Technology, Faculty of Health Sciences, Okayama University, Okayama, Japan
| | - Takao Hiraki
- Department of Radiology, Faculty of Medicine, Dentistry and Pharmaceutical Sciences, Okayama University, Okayama, Japan
| | - Hyungwoo Ahn
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
| |
Collapse
|