1
|
DeFoor MT, Sheean AJ. Editorial Commentary: Experts in Shoulder Surgery Do Not Consistently Detect Artificial Intelligence-Generated Scientific Abstracts. Arthroscopy 2024:S0749-8063(24)00641-8. [PMID: 39243996 DOI: 10.1016/j.arthro.2024.08.038] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/29/2024] [Accepted: 08/30/2024] [Indexed: 09/09/2024]
Abstract
There has been exponential growth in the number of artificial intelligence (AI)- and machine learning (ML)-related publications in recent years. For example, in the field of shoulder and elbow surgery, there was a 6-fold increase in the number of publications between 2018 and 2021. AI shows the potential to improve diagnostic precision, generate precise surgical templates, direct personalized treatment plans, and reduce administrative costs. However, although AI and ML technology has the ability to positively impact biomedical research, it should be closely monitored and used with extreme caution in the realm of research and scientific writing. Current large language models raise concerns regarding the veracity of AI-generated content, copyright and ownership infringement, fabricated references, lack of in-text citations, plagiarism, and questions of authorship. Recent research has shown that even the most experienced surgeons are unable to consistently detect AI-generated scientific writing. Of note, AI detection software is more adept in this role. AI should be used with caution in the development and production of scholarly work.
Collapse
Affiliation(s)
- Mikalyn T DeFoor
- San Antonio Military Medical Center, Ft Sam Houston, Texas, U.S.A
| | - Andrew J Sheean
- San Antonio Military Medical Center, Ft Sam Houston, Texas, U.S.A
| |
Collapse
|
2
|
Günay AE, Özer A, Yazıcı A, Sayer G. Comparison of ChatGPT versions in informing patients with rotator cuff injuries. JSES Int 2024; 8:1016-1018. [PMID: 39280147 PMCID: PMC11401580 DOI: 10.1016/j.jseint.2024.04.016] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/18/2024] Open
Abstract
Background The aim of this study is to evaluate whether Chat Generative Pretrained Transformer (ChatGPT) can be recommended as a resource for informing patients planning rotator cuff repairs, and to assess the differences between ChatGPT 3.5 and 4.0 versions in terms of information content and readability. Methods In August 2023, 13 commonly asked questions by patients with rotator cuff disease were posed to ChatGPT 3.5 and ChatGPT 4 programs using different internet protocol computers by 3 experienced surgeons in rotator cuff surgery. After converting the answers of both versions into text, the quality and readability of the answers were examined. Results The average Journal of the American Medical Association score for both versions was 0, and the average DISCERN score was 61.6. A statistically significant and strong correlation was found between ChatGPT 3.5 and 4.0 DISCERN scores. There was excellent agreement in DISCERN scores for both versions among the 3 evaluators. ChatGPT 3.5 was found to be less readable than ChatGPT 4.0. Conclusion The information provided by the ChatGPT conversational system was evaluated as of high quality, but there were significant shortcomings in terms of reliability due to the lack of citations. Despite the ChatGPT 4.0 version having higher readability scores, both versions were considered difficult to read.
Collapse
Affiliation(s)
- Ali Eray Günay
- Department of Orthopedics and Traumatology, Kayseri City Training and Research Hospital, Kayseri, Turkey
| | - Alper Özer
- Department of Orthopedics and Traumatology, Kayseri City Training and Research Hospital, Kayseri, Turkey
| | - Alparslan Yazıcı
- Department of Orthopedics and Traumatology, Develi State Hospital, Kayseri, Turkey
| | - Gökhan Sayer
- Department of Orthopedics and Traumatology, Bursa City Training and Research Hospital, Bursa, Turkey
| |
Collapse
|
3
|
Fatima A, Shafique MA, Alam K, Fadlalla Ahmed TK, Mustafa MS. ChatGPT in medicine: A cross-disciplinary systematic review of ChatGPT's (artificial intelligence) role in research, clinical practice, education, and patient interaction. Medicine (Baltimore) 2024; 103:e39250. [PMID: 39121303 PMCID: PMC11315549 DOI: 10.1097/md.0000000000039250] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/01/2024] [Accepted: 07/19/2024] [Indexed: 08/11/2024] Open
Abstract
BACKGROUND ChatGPT, a powerful AI language model, has gained increasing prominence in medicine, offering potential applications in healthcare, clinical decision support, patient communication, and medical research. This systematic review aims to comprehensively assess the applications of ChatGPT in healthcare education, research, writing, patient communication, and practice while also delineating potential limitations and areas for improvement. METHOD Our comprehensive database search retrieved relevant papers from PubMed, Medline and Scopus. After the screening process, 83 studies met the inclusion criteria. This review includes original studies comprising case reports, analytical studies, and editorials with original findings. RESULT ChatGPT is useful for scientific research and academic writing, and assists with grammar, clarity, and coherence. This helps non-English speakers and improves accessibility by breaking down linguistic barriers. However, its limitations include probable inaccuracy and ethical issues, such as bias and plagiarism. ChatGPT streamlines workflows and offers diagnostic and educational potential in healthcare but exhibits biases and lacks emotional sensitivity. It is useful in inpatient communication, but requires up-to-date data and faces concerns about the accuracy of information and hallucinatory responses. CONCLUSION Given the potential for ChatGPT to transform healthcare education, research, and practice, it is essential to approach its adoption in these areas with caution due to its inherent limitations.
Collapse
Affiliation(s)
- Afia Fatima
- Department of Medicine, Jinnah Sindh Medical University, Karachi, Pakistan
| | | | - Khadija Alam
- Department of Medicine, Liaquat National Medical College, Karachi, Pakistan
| | | | | |
Collapse
|
4
|
Large Language Models in Orthopaedic Publications: The Good, the Bad and the Ugly. Orthop J Sports Med 2024; 12:23259671241265705. [PMID: 39176267 PMCID: PMC11339935 DOI: 10.1177/23259671241265705] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 08/24/2024] Open
|
5
|
Rupp M, Moser LB, Hess S, Angele P, Aurich M, Dyrna F, Nehrer S, Neubauer M, Pawelczyk J, Izadpanah K, Zellner J, Niemeyer P. Orthopaedic surgeons display a positive outlook towards artificial intelligence: A survey among members of the AGA Society for Arthroscopy and Joint Surgery. J Exp Orthop 2024; 11:e12080. [PMID: 38974054 PMCID: PMC11227606 DOI: 10.1002/jeo2.12080] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/02/2024] [Revised: 06/13/2024] [Accepted: 06/21/2024] [Indexed: 07/09/2024] Open
Abstract
Purpose The purpose of this study was to evaluate the perspective of orthopaedic surgeons on the impact of artificial intelligence (AI) and to evaluate the influence of experience, workplace setting and familiarity with digital solutions on views on AI. Methods Orthopaedic surgeons of the AGA Society for Arthroscopy and Joint Surgery were invited to participate in an online, cross-sectional survey designed to gather information on professional background, subjective AI knowledge, opinion on the future impact of AI, openness towards different applications of AI, and perceived advantages and disadvantages of AI. Subgroup analyses were performed to examine the influence of experience, workplace setting and openness towards digital solutions on perspectives towards AI. Results Overall, 360 orthopaedic surgeons participated. The majority indicated average (43.6%) or rudimentary (38.1%) AI knowledge. Most (54.5%) expected AI to substantially influence orthopaedics within 5-10 years, predominantly as a complementary tool (91.1%). Preoperative planning (83.8%) was identified as the most likely clinical use case. A lack of consensus was observed regarding acceptable error levels. Time savings in preoperative planning (62.5%) and improved documentation (81%) were identified as notable advantages while declining skills of the next generation (64.5%) were rated as the most substantial drawback. There were significant differences in subjective AI knowledge depending on participants' experience (p = 0.021) and familiarity with digital solutions (p < 0.001), acceptable error levels depending on workplace setting (p = 0.004), and prediction of AI impact depending on familiarity with digital solutions (p < 0.001). Conclusion The majority of orthopaedic surgeons in this survey anticipated a notable positive impact of AI on their field, primarily as an assistive technology. A lack of consensus on acceptable error levels of AI and concerns about declining skills among future surgeons were observed. Level of Evidence Level IV, cross-sectional study.
Collapse
Affiliation(s)
- Marco‐Christopher Rupp
- Sektion Sportorthopädie, Klinikum rechts der IsarTechnische Universität MünchenMunichGermany
- Steadman Philippon Research InstituteVailColoradoUSA
| | - Lukas B. Moser
- Klinische Abteilung für Orthopädie und TraumatologieUniversitätsklinikum KremsKrems an der DonauAustria
- Zentrum für Regenerative MedizinUniversität für Weiterbildung KremsKrems an der DonauAustria
- SporthopaedicumRegensburgGermany
| | - Silvan Hess
- Universitätsklinik für Orthopädische Chirurgie und Traumatologie, InselspitalBernSwitzerland
| | - Peter Angele
- SporthopaedicumRegensburgGermany
- Klinik für Unfall‐ und WiederherstellungschirurgieUniversitätsklinikum RegensburgRegensburgGermany
| | | | | | - Stefan Nehrer
- Klinische Abteilung für Orthopädie und TraumatologieUniversitätsklinikum KremsKrems an der DonauAustria
- Zentrum für Regenerative MedizinUniversität für Weiterbildung KremsKrems an der DonauAustria
- Fakultät für Gesundheit und MedizinUniversität für Weiterbildung KremsKrems an der DonauAustria
| | - Markus Neubauer
- Klinische Abteilung für Orthopädie und TraumatologieUniversitätsklinikum KremsKrems an der DonauAustria
- Zentrum für Regenerative MedizinUniversität für Weiterbildung KremsKrems an der DonauAustria
| | - Johannes Pawelczyk
- Sektion Sportorthopädie, Klinikum rechts der IsarTechnische Universität MünchenMunichGermany
| | - Kaywan Izadpanah
- Klinik für Orthopädie und Unfallchirurgie, Universitätsklinikum Freiburg, Medizinische FakultätAlbert‐Ludwigs‐Universität FreiburgFreiburgGermany
| | | | - Philipp Niemeyer
- OCM – Orthopädische Chirurgie MünchenMunichGermany
- Albert‐Ludwigs‐UniversityFreiburgGermany
| | | |
Collapse
|
6
|
Wascher DC, Ollivier M. Large Language Models in Orthopaedic Publications: The Good, the Bad and the Ugly. Am J Sports Med 2024; 52:2193-2195. [PMID: 39101739 DOI: 10.1177/03635465241265692] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 08/06/2024]
|
7
|
Morya VK, Lee HW, Shahid H, Magar AG, Lee JH, Kim JH, Jun L, Noh KC. Application of ChatGPT for Orthopedic Surgeries and Patient Care. Clin Orthop Surg 2024; 16:347-356. [PMID: 38827766 PMCID: PMC11130626 DOI: 10.4055/cios23181] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/02/2023] [Revised: 11/15/2023] [Accepted: 12/12/2023] [Indexed: 06/05/2024] Open
Abstract
Artificial intelligence (AI) has rapidly transformed various aspects of life, and the launch of the chatbot "ChatGPT" by OpenAI in November 2022 has garnered significant attention and user appreciation. ChatGPT utilizes natural language processing based on a "generative pre-trained transfer" (GPT) model, specifically the transformer architecture, to generate human-like responses to a wide range of questions and topics. Equipped with approximately 57 billion words and 175 billion parameters from online data, ChatGPT has potential applications in medicine and orthopedics. One of its key strengths is its personalized, easy-to-understand, and adaptive response, which allows it to learn continuously through user interaction. This article discusses how AI, especially ChatGPT, presents numerous opportunities in orthopedics, ranging from preoperative planning and surgical techniques to patient education and medical support. Although ChatGPT's user-friendly responses and adaptive capabilities are laudable, its limitations, including biased responses and ethical concerns, necessitate its cautious and responsible use. Surgeons and healthcare providers should leverage the strengths of the ChatGPT while recognizing its current limitations and verifying critical information through independent research and expert opinions. As AI technology continues to evolve, ChatGPT may become a valuable tool in orthopedic education and patient care, leading to improved outcomes and efficiency in healthcare delivery. The integration of AI into orthopedics offers substantial benefits but requires careful consideration and continuous improvement.
Collapse
Affiliation(s)
- Vivek Kumar Morya
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Ho-Won Lee
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Hamzah Shahid
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Anuja Gajanan Magar
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Ju-Hyung Lee
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Jae-Hyung Kim
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Lang Jun
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| | - Kyu-Cheol Noh
- Department of Orthopedic Surgery, Hallym University Kangnam Sacred Heart Hospital, Seoul, Korea
| |
Collapse
|
8
|
Buldur M, Sezer B. Evaluating the accuracy of Chat Generative Pre-trained Transformer version 4 (ChatGPT-4) responses to United States Food and Drug Administration (FDA) frequently asked questions about dental amalgam. BMC Oral Health 2024; 24:605. [PMID: 38789962 PMCID: PMC11127407 DOI: 10.1186/s12903-024-04358-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2023] [Accepted: 05/09/2024] [Indexed: 05/26/2024] Open
Abstract
BACKGROUND The use of artificial intelligence in the field of health sciences is becoming widespread. It is known that patients benefit from artificial intelligence applications on various health issues, especially after the pandemic period. One of the most important issues in this regard is the accuracy of the information provided by artificial intelligence applications. OBJECTIVE The purpose of this study was to the frequently asked questions about dental amalgam, as determined by the United States Food and Drug Administration (FDA), which is one of these information resources, to Chat Generative Pre-trained Transformer version 4 (ChatGPT-4) and to compare the content of the answers given by the application with the answers of the FDA. METHODS The questions were directed to ChatGPT-4 on May 8th and May 16th, 2023, and the responses were recorded and compared at the word and meaning levels using ChatGPT. The answers from the FDA webpage were also recorded. The responses were compared for content similarity in "Main Idea", "Quality Analysis", "Common Ideas", and "Inconsistent Ideas" between ChatGPT-4's responses and FDA's responses. RESULTS ChatGPT-4 provided similar responses at one-week intervals. In comparison with FDA guidance, it provided answers with similar information content to frequently asked questions. However, although there were some similarities in the general aspects of the recommendation regarding amalgam removal in the question, the two texts are not the same, and they offered different perspectives on the replacement of fillings. CONCLUSIONS The findings of this study indicate that ChatGPT-4, an artificial intelligence based application, encompasses current and accurate information regarding dental amalgam and its removal, providing it to individuals seeking access to such information. Nevertheless, we believe that numerous studies are required to assess the validity and reliability of ChatGPT-4 across diverse subjects.
Collapse
Affiliation(s)
- Mehmet Buldur
- Department of Restorative Dentistry, School of Dentistry, Çanakkale Onsekiz Mart University, Çanakkale, Türkiye
| | - Berkant Sezer
- Department of Pediatric Dentistry, School of Dentistry, Çanakkale Onsekiz Mart University, Çanakkale, Türkiye.
| |
Collapse
|
9
|
Zhang S, Liau ZQG, Tan KLM, Chua WL. Evaluating the accuracy and relevance of ChatGPT responses to frequently asked questions regarding total knee replacement. Knee Surg Relat Res 2024; 36:15. [PMID: 38566254 PMCID: PMC10986046 DOI: 10.1186/s43019-024-00218-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/02/2024] [Accepted: 03/12/2024] [Indexed: 04/04/2024] Open
Abstract
BACKGROUND Chat Generative Pretrained Transformer (ChatGPT), a generative artificial intelligence chatbot, may have broad applications in healthcare delivery and patient education due to its ability to provide human-like responses to a wide range of patient queries. However, there is limited evidence regarding its ability to provide reliable and useful information on orthopaedic procedures. This study seeks to evaluate the accuracy and relevance of responses provided by ChatGPT to frequently asked questions (FAQs) regarding total knee replacement (TKR). METHODS A list of 50 clinically-relevant FAQs regarding TKR was collated. Each question was individually entered as a prompt to ChatGPT (version 3.5), and the first response generated was recorded. Responses were then reviewed by two independent orthopaedic surgeons and graded on a Likert scale for their factual accuracy and relevance. These responses were then classified into accurate versus inaccurate and relevant versus irrelevant responses using preset thresholds on the Likert scale. RESULTS Most responses were accurate, while all responses were relevant. Of the 50 FAQs, 44/50 (88%) of ChatGPT responses were classified as accurate, achieving a mean Likert grade of 4.6/5 for factual accuracy. On the other hand, 50/50 (100%) of responses were classified as relevant, achieving a mean Likert grade of 4.9/5 for relevance. CONCLUSION ChatGPT performed well in providing accurate and relevant responses to FAQs regarding TKR, demonstrating great potential as a tool for patient education. However, it is not infallible and can occasionally provide inaccurate medical information. Patients and clinicians intending to utilize this technology should be mindful of its limitations and ensure adequate supervision and verification of information provided.
Collapse
Affiliation(s)
- Siyuan Zhang
- Department of Orthopaedic Surgery, National University Health System, Level 11, NUHS Tower Block, 1E Kent Ridge Road, Singapore, 119228, Singapore.
| | - Zi Qiang Glen Liau
- Department of Orthopaedic Surgery, National University Health System, Level 11, NUHS Tower Block, 1E Kent Ridge Road, Singapore, 119228, Singapore
| | - Kian Loong Melvin Tan
- Department of Orthopaedic Surgery, National University Health System, Level 11, NUHS Tower Block, 1E Kent Ridge Road, Singapore, 119228, Singapore
| | - Wei Liang Chua
- Department of Orthopaedic Surgery, National University Health System, Level 11, NUHS Tower Block, 1E Kent Ridge Road, Singapore, 119228, Singapore
| |
Collapse
|
10
|
Sánchez-Rosenberg G, Magnéli M, Barle N, Kontakis MG, Müller AM, Wittauer M, Gordon M, Brodén C. ChatGPT-4 generates orthopedic discharge documents faster than humans maintaining comparable quality: a pilot study of 6 cases. Acta Orthop 2024; 95:152-156. [PMID: 38597205 PMCID: PMC10959013 DOI: 10.2340/17453674.2024.40182] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/01/2023] [Accepted: 01/28/2024] [Indexed: 04/11/2024] Open
Abstract
BACKGROUND AND PURPOSE Large language models like ChatGPT-4 have emerged. They hold the potential to reduce the administrative burden by generating everyday clinical documents, thus allowing the physician to spend more time with the patient. We aimed to assess both the quality and efficiency of discharge documents generated by ChatGPT-4 in comparison with those produced by physicians. PATIENTS AND METHODS To emulate real-world situations, the health records of 6 fictional orthopedic cases were created. Discharge documents for each case were generated by a junior attending orthopedic surgeon and an advanced orthopedic resident. ChatGPT-4 was then prompted to generate the discharge documents using the same health record information. The quality assessment was performed by an expert panel (n = 15) blinded to the source of the documents. As secondary outcome, the time required to generate the documents was compared, logging the duration of the creation of the discharge documents by the physician and by ChatGPT-4. RESULTS Overall, both ChatGPT-4 and physician-generated notes were comparable in quality. Notably, ChatGPT-4 generated discharge documents 10 times faster than the traditional method. 4 events of hallucinations were found in the ChatGPT-4-generated content, compared with 6 events in the human/physician produced notes. CONCLUSION ChatGPT-4 creates orthopedic discharge notes faster than physicians, with comparable quality. This shows it has great potential for making these documents more efficient in orthopedic care. ChatGPT-4 has the potential to significantly reduce the administrative burden on healthcare professionals.
Collapse
Affiliation(s)
| | - Martin Magnéli
- Karolinska Institute, Department of Clinical Sciences at Danderyd Hospital, Stockholm; Sweden
| | - Niklas Barle
- Karolinska Institute, Department of Clinical Sciences at Danderyd Hospital, Stockholm; Sweden
| | - Michael G Kontakis
- Department of Surgical Sciences, Orthopedics, Uppsala University Hospital, Uppsala, Sweden
| | - Andreas Marc Müller
- Department of Orthopedic and Trauma Surgery, University Hospital Basel, Switzerland
| | - Matthias Wittauer
- Department of Orthopedic and Trauma Surgery, University Hospital Basel, Switzerland
| | - Max Gordon
- Karolinska Institute, Department of Clinical Sciences at Danderyd Hospital, Stockholm; Sweden
| | - Cyrus Brodén
- Department of Surgical Sciences, Orthopedics, Uppsala University Hospital, Uppsala, Sweden.
| |
Collapse
|
11
|
Abi-Rafeh J, Xu HH, Kazan R, Tevlin R, Furnas H. Large Language Models and Artificial Intelligence: A Primer for Plastic Surgeons on the Demonstrated and Potential Applications, Promises, and Limitations of ChatGPT. Aesthet Surg J 2024; 44:329-343. [PMID: 37562022 DOI: 10.1093/asj/sjad260] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2023] [Revised: 08/02/2023] [Accepted: 08/04/2023] [Indexed: 08/12/2023] Open
Abstract
BACKGROUND The rapidly evolving field of artificial intelligence (AI) holds great potential for plastic surgeons. ChatGPT, a recently released AI large language model (LLM), promises applications across many disciplines, including healthcare. OBJECTIVES The aim of this article was to provide a primer for plastic surgeons on AI, LLM, and ChatGPT, including an analysis of current demonstrated and proposed clinical applications. METHODS A systematic review was performed identifying medical and surgical literature on ChatGPT's proposed clinical applications. Variables assessed included applications investigated, command tasks provided, user input information, AI-emulated human skills, output validation, and reported limitations. RESULTS The analysis included 175 articles reporting on 13 plastic surgery applications and 116 additional clinical applications, categorized by field and purpose. Thirty-four applications within plastic surgery are thus proposed, with relevance to different target audiences, including attending plastic surgeons (n = 17, 50%), trainees/educators (n = 8, 24.0%), researchers/scholars (n = 7, 21%), and patients (n = 2, 6%). The 15 identified limitations of ChatGPT were categorized by training data, algorithm, and ethical considerations. CONCLUSIONS Widespread use of ChatGPT in plastic surgery will depend on rigorous research of proposed applications to validate performance and address limitations. This systemic review aims to guide research, development, and regulation to safely adopt AI in plastic surgery.
Collapse
|
12
|
Christy M, Morris MT, Goldfarb CA, Dy CJ. Appropriateness and Reliability of an Online Artificial Intelligence Platform's Responses to Common Questions Regarding Distal Radius Fractures. J Hand Surg Am 2024; 49:91-98. [PMID: 38069953 DOI: 10.1016/j.jhsa.2023.10.019] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/19/2023] [Revised: 10/25/2023] [Accepted: 10/26/2023] [Indexed: 02/05/2024]
Abstract
PURPOSE Chat Generative Pre-Trained Transformer (ChatGPT) is a novel artificial intelligence chatbot that is changing the way humans gather information online. The purpose of this study was to investigate ChatGPT's ability to appropriately and reliably answer common questions regarding distal radius fractures. METHODS Thirty common questions regarding distal radius fractures were presented in an identical manner to the online ChatGPT-3.5 interface three separate times, yielding 90 unique responses because ChatGPT produces an original answer with each query. All responses were graded as "appropriate," "appropriate but incomplete," or "inappropriate" by a consensus discussion among three hand surgeon reviewers. The questions were additionally subcategorized into one of four domains based on Bloom's cognitive learning taxonomy, and descriptive statistics were reported. RESULTS Seventy of the 90 total responses (78%) produced by ChatGPT were "appropriate," and 29 of the 30 questions (97%) had at least one response considered appropriate (of the three possible). However, only 17 of the 30 questions (57%) were answered appropriately on all three iterations. The test-retest reliability of ChatGPT was poor with an intraclass correlation coefficient of 0.12. Finally, ChatGPT performed best answering questions requiring lower-order thinking skills (Bloom's levels 1-3) and less well on level 4 questions. CONCLUSIONS This study found that although ChatGPT has the capability to answer common questions regarding distal radius fractures, caution should be taken before implementing its use, given ChatGPT's inconsistency in providing a complete and accurate response to the same question every time. CLINICAL RELEVANCE As the popularity and technology of ChatGPT continue to grow, it is important to understand the potential and limitations of this platform to determine how it may be best implemented to improve patient care.
Collapse
Affiliation(s)
- Michele Christy
- Department of Orthopaedic Surgery, Washington University in St. Louis, St. Louis, MO
| | - Marie T Morris
- Department of Orthopaedic Surgery, Washington University in St. Louis, St. Louis, MO
| | - Charles A Goldfarb
- Department of Orthopaedic Surgery, Washington University in St. Louis, St. Louis, MO
| | - Christopher J Dy
- Department of Orthopaedic Surgery, Washington University in St. Louis, St. Louis, MO.
| |
Collapse
|
13
|
Liu JW, McCulloch PC. SONNET #29888: ChatGPT Finds Poetry in Anterior Cruciate Ligament Reconstruction and Return to Sport. Arthroscopy 2024; 40:197-198. [PMID: 38296427 DOI: 10.1016/j.arthro.2023.09.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/20/2023] [Accepted: 09/26/2023] [Indexed: 02/08/2024]
Affiliation(s)
- Jennifer W Liu
- Department of Orthopedic Surgery & Sports Medicine, Houston Methodist Hospital, Houston, Texas, U.S.A
| | - Patrick C McCulloch
- Department of Orthopedic Surgery & Sports Medicine, Houston Methodist Hospital, Houston, Texas, U.S.A
| |
Collapse
|
14
|
Ganjavi C, Eppler MB, Pekcan A, Biedermann B, Abreu A, Collins GS, Gill IS, Cacciamani GE. Publishers' and journals' instructions to authors on use of generative artificial intelligence in academic and scientific publishing: bibliometric analysis. BMJ 2024; 384:e077192. [PMID: 38296328 PMCID: PMC10828852 DOI: 10.1136/bmj-2023-077192] [Citation(s) in RCA: 13] [Impact Index Per Article: 13.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 11/29/2023] [Indexed: 02/05/2024]
Abstract
OBJECTIVES To determine the extent and content of academic publishers' and scientific journals' guidance for authors on the use of generative artificial intelligence (GAI). DESIGN Cross sectional, bibliometric study. SETTING Websites of academic publishers and scientific journals, screened on 19-20 May 2023, with the search updated on 8-9 October 2023. PARTICIPANTS Top 100 largest academic publishers and top 100 highly ranked scientific journals, regardless of subject, language, or country of origin. Publishers were identified by the total number of journals in their portfolio, and journals were identified through the Scimago journal rank using the Hirsch index (H index) as an indicator of journal productivity and impact. MAIN OUTCOME MEASURES The primary outcomes were the content of GAI guidelines listed on the websites of the top 100 academic publishers and scientific journals, and the consistency of guidance between the publishers and their affiliated journals. RESULTS Among the top 100 largest publishers, 24% provided guidance on the use of GAI, of which 15 (63%) were among the top 25 publishers. Among the top 100 highly ranked journals, 87% provided guidance on GAI. Of the publishers and journals with guidelines, the inclusion of GAI as an author was prohibited in 96% and 98%, respectively. Only one journal (1%) explicitly prohibited the use of GAI in the generation of a manuscript, and two (8%) publishers and 19 (22%) journals indicated that their guidelines exclusively applied to the writing process. When disclosing the use of GAI, 75% of publishers and 43% of journals included specific disclosure criteria. Where to disclose the use of GAI varied, including in the methods or acknowledgments, in the cover letter, or in a new section. Variability was also found in how to access GAI guidelines shared between journals and publishers. GAI guidelines in 12 journals directly conflicted with those developed by the publishers. The guidelines developed by top medical journals were broadly similar to those of academic journals. CONCLUSIONS Guidelines by some top publishers and journals on the use of GAI by authors are lacking. Among those that provided guidelines, the allowable uses of GAI and how it should be disclosed varied substantially, with this heterogeneity persisting in some instances among affiliated publishers and journals. Lack of standardization places a burden on authors and could limit the effectiveness of the regulations. As GAI continues to grow in popularity, standardized guidelines to protect the integrity of scientific output are needed.
Collapse
Affiliation(s)
- Conner Ganjavi
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| | - Michael B Eppler
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| | - Asli Pekcan
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| | - Brett Biedermann
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| | - Andre Abreu
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| | - Gary S Collins
- UK EQUATOR Centre, Centre for Statistics in Medicine, Nuffield Department of Orthopaedics, Rheumatology and Musculoskeletal Sciences, University of Oxford, Oxford, UK
| | - Inderbir S Gill
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| | - Giovanni E Cacciamani
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| |
Collapse
|
15
|
Younis HA, Eisa TAE, Nasser M, Sahib TM, Noor AA, Alyasiri OM, Salisu S, Hayder IM, Younis HA. A Systematic Review and Meta-Analysis of Artificial Intelligence Tools in Medicine and Healthcare: Applications, Considerations, Limitations, Motivation and Challenges. Diagnostics (Basel) 2024; 14:109. [PMID: 38201418 PMCID: PMC10802884 DOI: 10.3390/diagnostics14010109] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/29/2023] [Revised: 12/02/2023] [Accepted: 12/04/2023] [Indexed: 01/12/2024] Open
Abstract
Artificial intelligence (AI) has emerged as a transformative force in various sectors, including medicine and healthcare. Large language models like ChatGPT showcase AI's potential by generating human-like text through prompts. ChatGPT's adaptability holds promise for reshaping medical practices, improving patient care, and enhancing interactions among healthcare professionals, patients, and data. In pandemic management, ChatGPT rapidly disseminates vital information. It serves as a virtual assistant in surgical consultations, aids dental practices, simplifies medical education, and aids in disease diagnosis. A total of 82 papers were categorised into eight major areas, which are G1: treatment and medicine, G2: buildings and equipment, G3: parts of the human body and areas of the disease, G4: patients, G5: citizens, G6: cellular imaging, radiology, pulse and medical images, G7: doctors and nurses, and G8: tools, devices and administration. Balancing AI's role with human judgment remains a challenge. A systematic literature review using the PRISMA approach explored AI's transformative potential in healthcare, highlighting ChatGPT's versatile applications, limitations, motivation, and challenges. In conclusion, ChatGPT's diverse medical applications demonstrate its potential for innovation, serving as a valuable resource for students, academics, and researchers in healthcare. Additionally, this study serves as a guide, assisting students, academics, and researchers in the field of medicine and healthcare alike.
Collapse
Affiliation(s)
- Hussain A. Younis
- College of Education for Women, University of Basrah, Basrah 61004, Iraq
| | | | - Maged Nasser
- Computer & Information Sciences Department, Universiti Teknologi PETRONAS, Seri Iskandar 32610, Malaysia;
| | - Thaeer Mueen Sahib
- Kufa Technical Institute, Al-Furat Al-Awsat Technical University, Kufa 54001, Iraq;
| | - Ameen A. Noor
- Computer Science Department, College of Education, University of Almustansirya, Baghdad 10045, Iraq;
| | | | - Sani Salisu
- Department of Information Technology, Federal University Dutse, Dutse 720101, Nigeria;
| | - Israa M. Hayder
- Qurna Technique Institute, Southern Technical University, Basrah 61016, Iraq;
| | - Hameed AbdulKareem Younis
- Department of Cybersecurity, College of Computer Science and Information Technology, University of Basrah, Basrah 61016, Iraq;
| |
Collapse
|
16
|
Kayaalp ME, Ollivier M, Winkler PW, Dahmen J, Musahl V, Hirschmann MT, Karlsson J. Embrace responsible ChatGPT usage to overcome language barriers in academic writing. Knee Surg Sports Traumatol Arthrosc 2024; 32:5-9. [PMID: 38226673 DOI: 10.1002/ksa.12014] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/21/2023] [Accepted: 11/08/2023] [Indexed: 01/17/2024]
Affiliation(s)
- M Enes Kayaalp
- Department of Orthopaedic Surgery, University of Pittsburgh, Pittsburgh, Pennsylvania, USA
- Department for Orthopaedics and Traumatology, Istanbul Kartal Research and Training Hospital, Istanbul, Turkiye
| | - Matthieu Ollivier
- CNRS, Institute of Movement Sciences (ISM), Aix Marseille University, Marseille, France
| | - Philipp W Winkler
- Department for Orthopaedics and Traumatology, Kepler University Hospital GmbH, Linz, Austria
| | - Jari Dahmen
- Department of Orthopaedic Surgery and Sports Medicine, Amsterdam Movement Sciences, Amsterdam UMC, University of Amsterdam, Amsterdam, The Netherlands
- Academic Center for Evidence Based Sports Medicine (ACES), Amsterdam, The Netherlands
- Amsterdam Collaboration for Health and Safety in Sports (ACHSS), International Olympic Committee (IOC) Research Center Amsterdam UMC, Amsterdam, The Netherlands
| | - Volker Musahl
- Department of Orthopaedic Surgery, University of Pittsburgh, Pittsburgh, Pennsylvania, USA
| | - Michael T Hirschmann
- Department of Orthopedic Surgery and Traumatology, Head Knee Surgery and DKF Head of Research, Kantonsspital Baselland, Bruderholz, Bottmingen, Switzerland
- University of Basel, Basel, Switzerland
| | - Jon Karlsson
- Department for Orthopaedics, Sahlgrenska University Hospital, Institute of Clinical Sciences, Sahlgrenska Academy, Gothenburg University, Gothenburg, Sweden
| |
Collapse
|
17
|
Kunze KN, Williams RJ, Ranawat AS, Pearle AD, Kelly BT, Karlsson J, Martin RK, Pareek A. Artificial intelligence (AI) and large data registries: Understanding the advantages and limitations of contemporary data sets for use in AI research. Knee Surg Sports Traumatol Arthrosc 2024; 32:13-18. [PMID: 38226678 DOI: 10.1002/ksa.12018] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/09/2023] [Accepted: 11/27/2023] [Indexed: 01/17/2024]
Affiliation(s)
- Kyle N Kunze
- Department of Orthopaedic Surgery, Hospital for Special Surgery, New York, New York, USA
- Sports Medicine and Shoulder Service, Hospital for Special Surgery, New York, New York, USA
| | - Riley J Williams
- Department of Orthopaedic Surgery, Hospital for Special Surgery, New York, New York, USA
- Sports Medicine and Shoulder Service, Hospital for Special Surgery, New York, New York, USA
| | - Anil S Ranawat
- Department of Orthopaedic Surgery, Hospital for Special Surgery, New York, New York, USA
- Sports Medicine and Shoulder Service, Hospital for Special Surgery, New York, New York, USA
| | - Andrew D Pearle
- Department of Orthopaedic Surgery, Hospital for Special Surgery, New York, New York, USA
- Sports Medicine and Shoulder Service, Hospital for Special Surgery, New York, New York, USA
| | - Bryan T Kelly
- Department of Orthopaedic Surgery, Hospital for Special Surgery, New York, New York, USA
- Sports Medicine and Shoulder Service, Hospital for Special Surgery, New York, New York, USA
| | - Jon Karlsson
- Department of Orthopaedics, Sahlgrenska University Hospital, Sahlgrenska Academy, Gothenburg University, Gothenburg, Sweden
| | - R Kyle Martin
- Department of Orthopedic Surgery, University of Minnesota, Minneapolis, Minnesota, USA
| | - Ayoosh Pareek
- Department of Orthopaedic Surgery, Hospital for Special Surgery, New York, New York, USA
- Sports Medicine and Shoulder Service, Hospital for Special Surgery, New York, New York, USA
| |
Collapse
|
18
|
Bisi T, Risser A, Clavert P, Migaud H, Dartus J. What is the rate of text generated by artificial intelligence over a year of publication in Orthopedics & Traumatology: Surgery & Research? Analysis of 425 articles before versus after the launch of ChatGPT in November 2022. Orthop Traumatol Surg Res 2023; 109:103694. [PMID: 37776949 DOI: 10.1016/j.otsr.2023.103694] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/17/2023] [Revised: 07/10/2023] [Accepted: 08/24/2023] [Indexed: 10/02/2023]
Abstract
BACKGROUND The use of artificial intelligence (AI) is soaring, and the launch of ChatGPT in November 2022 has accelerated this trend. This "chatbot" can generate complete scientific articles, with risk of plagiarism by mining existing data or downright fraud by fabricating studies with no real data at all. There are tools that detect AI in publications, but to our knowledge they have not been systematically assessed for publication in scientific journals. We therefore conducted a retrospective study on articles published in Orthopaedics & Traumatology: Surgery & Research (OTSR): firstly, to screen for AI-generated content before and after the publicized launch of ChatGPT; secondly, to assess whether AI was more often used in some countries than others to generate content; thirdly, to determine whether plagiarism rate correlated with AI-generation, and lastly, to determine whether elements other than text generation, and notably the translation procedure, could raise suspicion of AI use. HYPOTHESIS The rate of AI use increased after the publicized launch of ChatGPT v3.5 in November 2022. MATERIAL AND METHODS In all, 425 articles published between February 2022 and September 2023 (221 before and 204 after November 1, 2022) underwent ZeroGPT assessment of the level of AI generation in the final English-language version (abstract and body of the article). Two scores were obtained: probability of AI generation, in six grades from Human to AI; and percentage AI generation. Plagiarism was assessed on the Ithenticate application at submission. Articles in French were assessed in their English-language version as translated by a human translator, with comparison to automatic translation by Google Translate and DeepL. RESULTS AI-generated text was detected mainly in Abstracts, with a 10.1% rate of AI or considerable AI generation, compared to only 1.9% for the body of the article and 5.6% for the total body+abstract. Analysis for before and after November 2022 found an increase in AI generation in body+abstract, from 10.30±15.95% (range, 0-100%) to 15.64±19.8% (range, 0-99.93) (p < 0.04; NS for abstracts alone). AI scores differed between types of article: 14.9% for original articles and 9.8% for reviews (p<0.01). The highest rates of probable AI generation were in articles from Japan, China, South America and English-speaking countries (p<0.0001). Plagiarism rates did not increase between the two study periods, and were unrelated to AI rates. On the other hand, when articles were classified as "suspected" of AI generation (plagiarism rate ≥ 20%) or "non-suspected" (rate<20%), the "similarity" score was higher in suspect articles: 25.7±13.23% (range, 10-69%) versus 16.28±10% (range, 0-79%) (p < 0.001). In the body of the article, use of translation software was associated with higher AI rates than with a human translator: 3.5±5% for human translators, versus 18±10% and 21.9±11% respectively for Google Translate and DeepL (p < 0.001). DISCUSSION The present study revealed an increasing rate of AI use in articles published in OTSR. AI grades differed according to type of article and country of origin. Use of translation software increased the AI grade. In the long run, use of ChatGPT incurs a risk of plagiarism and scientific misconduct, and needs to be detected and signaled by a digital tag on any robot-generated text. LEVEL OF EVIDENCE III; case-control study.
Collapse
Affiliation(s)
- Théophile Bisi
- Département universitaire de chirurgie orthopédique, université de Lille, CHU de Lille, 59000 Lille, France; Service de chirurgie orthopédique, centre hospitalier universitaire (CHU) de Lille, hôpital Roger-Salengro, place de Verdun, 59000 Lille, France.
| | - Anthony Risser
- Service de chirurgie du membre supérieur, Hautepierre 2, CHRU Strasbourg, 1, avenue Molière, 67200 Strasbourg, France
| | - Philippe Clavert
- Service de chirurgie du membre supérieur, Hautepierre 2, CHRU Strasbourg, 1, avenue Molière, 67200 Strasbourg, France; Faculté de médecine, institut d'anatomie normale, 4, rue Kirschleger, 67085 Strasbourg, France
| | - Henri Migaud
- Département universitaire de chirurgie orthopédique, université de Lille, CHU de Lille, 59000 Lille, France; Service de chirurgie orthopédique, centre hospitalier universitaire (CHU) de Lille, hôpital Roger-Salengro, place de Verdun, 59000 Lille, France
| | - Julien Dartus
- Département universitaire de chirurgie orthopédique, université de Lille, CHU de Lille, 59000 Lille, France; Service de chirurgie orthopédique, centre hospitalier universitaire (CHU) de Lille, hôpital Roger-Salengro, place de Verdun, 59000 Lille, France
| |
Collapse
|
19
|
Maroteau G, An JS, Murgier J, Hulet C, Ollivier M, Ferreira A. Evaluation of the impact of large language learning models on articles submitted to Orthopaedics & Traumatology: Surgery & Research (OTSR): A significant increase in the use of artificial intelligence in 2023. Orthop Traumatol Surg Res 2023; 109:103720. [PMID: 37866509 DOI: 10.1016/j.otsr.2023.103720] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/04/2023] [Revised: 10/04/2023] [Accepted: 10/04/2023] [Indexed: 10/24/2023]
Abstract
INTRODUCTION There has been an unprecedented rise is the use of artificial intelligence (AI) amongst medical fields. Recently, a dialogue agent called ChatGPT (Generative Pre-trained Transformer) has grown in popularity through its use of large language models (LLM) to clearly and precisely generate text on demand. However, the impact of AI on the creation of scientific articles is remains unknown. A retrospective study was carried out with the aim of answering the following questions: identify the presence of text generated by LLM before and after the increased usage of ChatGPT in articles submitted in OTSR; determine if the type of article, the year of submission, and the country of origin, influenced the proportion of text generated, at least in part by AI. MATERIAL AND METHODS A total of 390 English articles were submitted to OTSR in January, February and March 2022 (n=204) and over the same months of 2023 (n=186) were analyzed. All articles were analyzed using the ZeroGPT tool, which provides an assumed rate of AI use expressed as a percentage. A comparison of the average rate of AI use was carried out between the articles submitted in 2022 and 2023. This comparison was repeated keeping only the articles with the highest percentage of suspected AI use (greater than 10 and 20%). A secondary analysis was carried out to identify risk factors for AI use. RESULTS The average percentage of suspected LLM use in the entire cohort was 11%±6, with 160 articles (41.0%) having a suspected AI rate greater than 10% and 61 (15.6%) with an assumed AI rate greater than 20%. A comparison between articles submitted in 2022 and 2023 revealed a significant increase in the use of these tools after the launch of ChatGPT 3.5 (9.4% in 2022 and 12.6% in 2023 [p=0.004]). The number of articles with suspected AI rates of greater than 10 and 20% were significantly higher in 2023: >10%: 71 articles (34.8%) versus 89 articles (47.8%) (p=0.008) and >20%: 21 articles (10.3%) versus 40 articles (21.5%) (p=0.002). A risk factor analysis for LLLM use, demonstrated that authors of Asian geographic origin, and the submission year 2023 were associated with a higher rate of suspected AI use. An AI rate >20% was associated to Asian geographical origin with an odds ratio of 1.79 (95% CI: 1.03-3.11) (p=0.029), while the year of submission being 2023 had an odds ratio of 1.7 (95% CI: 1.1-2.5) (p=0.02). CONCLUSION This study highlights a significant increase in the use of LLM in the writing of articles submitted to the OTSR journal after the launch of ChatGPT 3.5. The increasing use of these models raises questions about originality and plagiarism in scientific research. AI offers creative opportunities but also raises ethical and methodological challenges. LEVEL OF EVIDENCE III; case control study.
Collapse
Affiliation(s)
- Gaëlle Maroteau
- Unité Inserm Comète 1075, Department of Orthopaedics and Traumatology, Caen University Hospital, avenue Cote-de-Nacre, 14000 Caen, France
| | - Jae-Sung An
- Tokyo Medical and Dental University, 1 Chome-5-45 Yushima, Bunkyo City, Tokyo 113-8510, Japan
| | - Jérome Murgier
- Service de chirurgie orthopédique, clinique Aguiléra, 21, rue de l'Estagnas, 64200 Biarritz, France
| | - Christophe Hulet
- Unité Inserm Comète 1075, Department of Orthopaedics and Traumatology, Caen University Hospital, avenue Cote-de-Nacre, 14000 Caen, France
| | - Matthieu Ollivier
- Institute of Movement and Locomotion, Department of Orthopaedics and Traumatology, Sainte-Marguerite Hospital, BP 29, 270, boulevard Sainte-Marguerite, 13274 Marseille, France; Aix-Marseille Unit, Institute for Locomotion, Department of Orthopaedics and Traumatology, CNRS, ISM, Sainte-Marguerite Hospital, AP-HM, Marseille, France
| | - Alexandre Ferreira
- Unité Inserm Comète 1075, Department of Orthopaedics and Traumatology, Caen University Hospital, avenue Cote-de-Nacre, 14000 Caen, France.
| |
Collapse
|
20
|
Zsidai B, Hilkert AS, Kaarre J, Narup E, Senorski EH, Grassi A, Ley C, Longo UG, Herbst E, Hirschmann MT, Kopf S, Seil R, Tischer T, Samuelsson K, Feldt R. A practical guide to the implementation of AI in orthopaedic research - part 1: opportunities in clinical application and overcoming existing challenges. J Exp Orthop 2023; 10:117. [PMID: 37968370 PMCID: PMC10651597 DOI: 10.1186/s40634-023-00683-z] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/24/2023] [Accepted: 10/21/2023] [Indexed: 11/17/2023] Open
Abstract
Artificial intelligence (AI) has the potential to transform medical research by improving disease diagnosis, clinical decision-making, and outcome prediction. Despite the rapid adoption of AI and machine learning (ML) in other domains and industry, deployment in medical research and clinical practice poses several challenges due to the inherent characteristics and barriers of the healthcare sector. Therefore, researchers aiming to perform AI-intensive studies require a fundamental understanding of the key concepts, biases, and clinical safety concerns associated with the use of AI. Through the analysis of large, multimodal datasets, AI has the potential to revolutionize orthopaedic research, with new insights regarding the optimal diagnosis and management of patients affected musculoskeletal injury and disease. The article is the first in a series introducing fundamental concepts and best practices to guide healthcare professionals and researcher interested in performing AI-intensive orthopaedic research studies. The vast potential of AI in orthopaedics is illustrated through examples involving disease- or injury-specific outcome prediction, medical image analysis, clinical decision support systems and digital twin technology. Furthermore, it is essential to address the role of human involvement in training unbiased, generalizable AI models, their explainability in high-risk clinical settings and the implementation of expert oversight and clinical safety measures for failure. In conclusion, the opportunities and challenges of AI in medicine are presented to ensure the safe and ethical deployment of AI models for orthopaedic research and clinical application. Level of evidence IV.
Collapse
Affiliation(s)
- Bálint Zsidai
- Sahlgrenska Sports Medicine Center, Gothenburg, Sweden.
- Department of Orthopaedics, Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Gothenburg, Sweden.
| | - Ann-Sophie Hilkert
- Department of Computer Science and Engineering, Chalmers University of Technology, Gothenburg, Sweden
- Medfield Diagnostics AB, Gothenburg, Sweden
| | - Janina Kaarre
- Sahlgrenska Sports Medicine Center, Gothenburg, Sweden
- Department of Orthopaedics, Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Gothenburg, Sweden
- Department of Orthopaedic Surgery, UPMC Freddie Fu Sports Medicine Center, University of Pittsburgh, Pittsburgh, USA
| | - Eric Narup
- Sahlgrenska Sports Medicine Center, Gothenburg, Sweden
- Department of Orthopaedics, Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Gothenburg, Sweden
| | - Eric Hamrin Senorski
- Sahlgrenska Sports Medicine Center, Gothenburg, Sweden
- Department of Health and Rehabilitation, Institute of Neuroscience and Physiology, Sahlgrenska Academy, University of Gothenburg, Gothenburg, Sweden
- Sportrehab Sports Medicine Clinic, Gothenburg, Sweden
| | - Alberto Grassi
- Department of Orthopaedics, Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Gothenburg, Sweden
- IIa Clinica Ortopedica E Traumatologica, IRCCS Istituto Ortopedico Rizzoli, Bologna, Italy
| | - Christophe Ley
- Department of Mathematics, University of Luxembourg, Esch-Sur-Alzette, Luxembourg
| | - Umile Giuseppe Longo
- Department of Orthopaedic and Trauma Surgery, Campus Bio-Medico University, Rome, Italy
| | - Elmar Herbst
- Department of Trauma, Hand and Reconstructive Surgery, University Hospital Münster, Münster, Germany
| | - Michael T Hirschmann
- Department of Orthopedic Surgery and Traumatology, Head Knee Surgery and DKF Head of Research, Kantonsspital Baselland, 4101, Bruderholz, Switzerland
| | - Sebastian Kopf
- Center of Orthopaedics and Traumatology, University Hospital Brandenburg a.d.H., Brandenburg Medical School Theodor Fontane, 14770, Brandenburg a.d.H., Germany
- Faculty of Health Sciences Brandenburg, Brandenburg Medical School Theodor Fontane, 14770, Brandenburg a.d.H., Germany
| | - Romain Seil
- Department of Orthopaedic Surgery, Centre Hospitalier Luxembourg and Luxembourg Institute of Health, Luxembourg, Luxembourg
| | - Thomas Tischer
- Clinic for Orthopaedics and Trauma Surgery, Malteser Waldkrankenhaus St. Marien, Erlangen, Germany
| | - Kristian Samuelsson
- Sahlgrenska Sports Medicine Center, Gothenburg, Sweden
- Department of Orthopaedics, Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Gothenburg, Sweden
- Department of Orthopaedics, Sahlgrenska University Hospital, Mölndal, Sweden
| | - Robert Feldt
- Department of Orthopaedics, Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Gothenburg, Sweden
| |
Collapse
|
21
|
Giorgino R, Alessandri-Bonetti M, Luca A, Migliorini F, Rossi N, Peretti GM, Mangiavini L. ChatGPT in orthopedics: a narrative review exploring the potential of artificial intelligence in orthopedic practice. Front Surg 2023; 10:1284015. [PMID: 38026475 PMCID: PMC10654618 DOI: 10.3389/fsurg.2023.1284015] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/27/2023] [Accepted: 10/16/2023] [Indexed: 12/01/2023] Open
Abstract
The field of orthopedics faces complex challenges requiring quick and intricate decisions, with patient education and compliance playing crucial roles in treatment outcomes. Technological advancements in artificial intelligence (AI) can potentially enhance orthopedic care. ChatGPT, a natural language processing technology developed by OpenAI, has shown promise in various sectors, including healthcare. ChatGPT can facilitate patient information exchange in orthopedics, provide clinical decision support, and improve patient communication and education. It can assist in differential diagnosis, suggest appropriate imaging modalities, and optimize treatment plans based on evidence-based guidelines. However, ChatGPT has limitations, such as insufficient expertise in specialized domains and a lack of contextual understanding. The application of ChatGPT in orthopedics is still evolving, with studies exploring its potential in clinical decision-making, patient education, workflow optimization, and scientific literature. The results indicate both the benefits and limitations of ChatGPT, emphasizing the need for caution, ethical considerations, and human oversight. Addressing training data quality, biases, data privacy, and accountability challenges is crucial for responsible implementation. While ChatGPT has the potential to transform orthopedic healthcare, further research and development are necessary to ensure its reliability, accuracy, and ethical use in patient care.
Collapse
Affiliation(s)
- Riccardo Giorgino
- IRCCS Istituto Ortopedico Galeazzi, Milan, Italy
- Residency Program in Orthopedics and Traumatology, University of Milan, Milan, Italy
| | - Mario Alessandri-Bonetti
- Department of Plastic Surgery, University of Pittsburgh Medical Center, Pittsburgh, PA, United States
| | - Andrea Luca
- IRCCS Istituto Ortopedico Galeazzi, Milan, Italy
| | - Filippo Migliorini
- Department of Orthopaedic, Trauma, and Reconstructive Surgery, RWTH University Medical Centre, Aachen, Germany
- Department of Orthopedics and Trauma Surgery, Academic Hospital of Bolzano (SABES-ASDAA), Teaching Hospital of the Paracelsus Medical University, Bolzano, Italy
| | - Nicolò Rossi
- IRCCS Istituto Ortopedico Galeazzi, Milan, Italy
- Residency Program in Orthopedics and Traumatology, University of Milan, Milan, Italy
| | - Giuseppe M. Peretti
- IRCCS Istituto Ortopedico Galeazzi, Milan, Italy
- Dipartimento di Scienze Biomediche per la Salute, Università degli Studi di Milano, Milan, Italy
| | - Laura Mangiavini
- IRCCS Istituto Ortopedico Galeazzi, Milan, Italy
- Dipartimento di Scienze Biomediche per la Salute, Università degli Studi di Milano, Milan, Italy
| |
Collapse
|
22
|
Kaarre J, Feldt R, Keeling LE, Dadoo S, Zsidai B, Hughes JD, Samuelsson K, Musahl V. Exploring the potential of ChatGPT as a supplementary tool for providing orthopaedic information. Knee Surg Sports Traumatol Arthrosc 2023; 31:5190-5198. [PMID: 37553552 PMCID: PMC10598178 DOI: 10.1007/s00167-023-07529-2] [Citation(s) in RCA: 24] [Impact Index Per Article: 24.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/19/2023] [Accepted: 07/26/2023] [Indexed: 08/10/2023]
Abstract
PURPOSE To investigate the potential use of large language models (LLMs) in orthopaedics by presenting queries pertinent to anterior cruciate ligament (ACL) surgery to generative pre-trained transformer (ChatGPT, specifically using its GPT-4 model of March 14th 2023). Additionally, this study aimed to evaluate the depth of the LLM's knowledge and investigate its adaptability to different user groups. It was hypothesized that the ChatGPT would be able to adapt to different target groups due to its strong language understanding and processing capabilities. METHODS ChatGPT was presented with 20 questions and response was requested for two distinct target audiences: patients and non-orthopaedic medical doctors. Two board-certified orthopaedic sports medicine surgeons and two expert orthopaedic sports medicine surgeons independently evaluated the responses generated by ChatGPT. Mean correctness, completeness, and adaptability to the target audiences (patients and non-orthopaedic medical doctors) were determined. A three-point response scale facilitated nuanced assessment. RESULTS ChatGPT exhibited fair accuracy, with average correctness scores of 1.69 and 1.66 (on a scale from 0, incorrect, 1, partially correct, to 2, correct) for patients and medical doctors, respectively. Three of the 20 questions (15.0%) were deemed incorrect by any of the four orthopaedic sports medicine surgeon assessors. Moreover, overall completeness was calculated to be 1.51 and 1.64 for patients and medical doctors, respectively, while overall adaptiveness was determined to be 1.75 and 1.73 for patients and doctors, respectively. CONCLUSION Overall, ChatGPT was successful in generating correct responses in approximately 65% of the cases related to ACL surgery. The findings of this study imply that LLMs offer potential as a supplementary tool for acquiring orthopaedic knowledge. However, although ChatGPT can provide guidance and effectively adapt to diverse target audiences, it cannot supplant the expertise of orthopaedic sports medicine surgeons in diagnostic and treatment planning endeavours due to its limited understanding of orthopaedic domains and its potential for erroneous responses. LEVEL OF EVIDENCE V.
Collapse
Affiliation(s)
- Janina Kaarre
- Department of Orthopaedic Surgery, UPMC Freddie Fu Sports Medicine Center, University of Pittsburgh, Pittsburgh, USA
- Department of Orthopaedics, Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Göteborgsvägen 31, 431 80 Mölndal, Sweden
| | - Robert Feldt
- Department of Computer Science and Engineering, Chalmers University of Technology, Gothenburg, Sweden
| | - Laura E. Keeling
- Department of Orthopaedic Surgery, UPMC Freddie Fu Sports Medicine Center, University of Pittsburgh, Pittsburgh, USA
| | - Sahil Dadoo
- Department of Orthopaedic Surgery, UPMC Freddie Fu Sports Medicine Center, University of Pittsburgh, Pittsburgh, USA
| | - Bálint Zsidai
- Department of Orthopaedics, Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Göteborgsvägen 31, 431 80 Mölndal, Sweden
| | - Jonathan D. Hughes
- Department of Orthopaedic Surgery, UPMC Freddie Fu Sports Medicine Center, University of Pittsburgh, Pittsburgh, USA
| | - Kristian Samuelsson
- Department of Orthopaedics, Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Göteborgsvägen 31, 431 80 Mölndal, Sweden
- Department of Orthopaedics, Sahlgrenska University Hospital, Mölndal, Sweden
| | - Volker Musahl
- Department of Orthopaedic Surgery, UPMC Freddie Fu Sports Medicine Center, University of Pittsburgh, Pittsburgh, USA
| |
Collapse
|
23
|
Deng L, Zhang Y, Luo S, Xu J. GPT-4 in breast cancer combat: a dazzling leap forward or merely a whim? Int J Surg 2023; 109:3732-3735. [PMID: 37994733 PMCID: PMC10651236 DOI: 10.1097/js9.0000000000000668] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2023] [Accepted: 07/30/2023] [Indexed: 11/24/2023]
Affiliation(s)
- Linfang Deng
- Department of Nursing, Jinzhou Medical University
| | - Yang Zhang
- Department of Breast Surgery, Xingtai People’s Hospital of Hebei Medical University, Xingtai, Hebei, People’s Republic of China
| | - Shaoting Luo
- Department of Pediatric Orthopedics, Shengjing Hospital of China Medical University, Shenyang, Liaoning
| | - Jinjiang Xu
- Department of Health Management Center, The First Hospital of Jinzhou Medical University, Jinzhou
| |
Collapse
|
24
|
Ghanem D, Covarrubias O, Raad M, LaPorte D, Shafiq B. ChatGPT Performs at the Level of a Third-Year Orthopaedic Surgery Resident on the Orthopaedic In-Training Examination. JB JS Open Access 2023; 8:e23.00103. [PMID: 38638869 PMCID: PMC11025881 DOI: 10.2106/jbjs.oa.23.00103] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 04/20/2024] Open
Abstract
Introduction Publicly available AI language models such as ChatGPT have demonstrated utility in text generation and even problem-solving when provided with clear instructions. Amidst this transformative shift, the aim of this study is to assess ChatGPT's performance on the orthopaedic surgery in-training examination (OITE). Methods All 213 OITE 2021 web-based questions were retrieved from the AAOS-ResStudy website (https://www.aaos.org/education/examinations/ResStudy). Two independent reviewers copied and pasted the questions and response options into ChatGPT Plus (version 4.0) and recorded the generated answers. All media-containing questions were flagged and carefully examined. Twelve OITE media-containing questions that relied purely on images (clinical pictures, radiographs, MRIs, CT scans) and could not be rationalized from the clinical presentation were excluded. Cohen's Kappa coefficient was used to examine the agreement of ChatGPT-generated responses between reviewers. Descriptive statistics were used to summarize the performance (% correct) of ChatGPT Plus. The 2021 norm table was used to compare ChatGPT Plus' performance on the OITE to national orthopaedic surgery residents in that same year. Results A total of 201 questions were evaluated by ChatGPT Plus. Excellent agreement was observed between raters for the 201 ChatGPT-generated responses, with a Cohen's Kappa coefficient of 0.947. 45.8% (92/201) were media-containing questions. ChatGPT had an average overall score of 61.2% (123/201). Its score was 64.2% (70/109) on non-media questions. When compared to the performance of all national orthopaedic surgery residents in 2021, ChatGPT Plus performed at the level of an average PGY3. Discussion ChatGPT Plus is able to pass the OITE with an overall score of 61.2%, ranking at the level of a third-year orthopaedic surgery resident. It provided logical reasoning and justifications that may help residents improve their understanding of OITE cases and general orthopaedic principles. Further studies are still needed to examine their efficacy and impact on long-term learning and OITE/ABOS performance.
Collapse
Affiliation(s)
- Diane Ghanem
- Department of Orthopaedic Surgery, The Johns Hopkins Hospital, Baltimore, Maryland
| | - Oscar Covarrubias
- School of Medicine, The Johns Hopkins University, Baltimore, Maryland
| | - Micheal Raad
- Department of Orthopaedic Surgery, The Johns Hopkins Hospital, Baltimore, Maryland
| | - Dawn LaPorte
- Department of Orthopaedic Surgery, The Johns Hopkins Hospital, Baltimore, Maryland
| | - Babar Shafiq
- Department of Orthopaedic Surgery, The Johns Hopkins Hospital, Baltimore, Maryland
| |
Collapse
|
25
|
Ray PP. A critical analysis of use of ChatGPT in orthopaedics. INTERNATIONAL ORTHOPAEDICS 2023; 47:2617-2618. [PMID: 37515601 DOI: 10.1007/s00264-023-05916-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/18/2023] [Accepted: 07/21/2023] [Indexed: 07/31/2023]
Affiliation(s)
- Partha Pratim Ray
- Department of Computer Applications, Sikkim University, 6th Mile, PO-Tadong, Gangtok, Sikkim, 737102, India.
| |
Collapse
|
26
|
Anastasio AT, Mills FB, Karavan MP, Adams SB. Evaluating the Quality and Usability of Artificial Intelligence-Generated Responses to Common Patient Questions in Foot and Ankle Surgery. FOOT & ANKLE ORTHOPAEDICS 2023; 8:24730114231209919. [PMID: 38027458 PMCID: PMC10666700 DOI: 10.1177/24730114231209919] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2023] Open
Abstract
Background Artificial intelligence (AI) platforms, such as ChatGPT, have become increasingly popular outlets for the consumption and distribution of health care-related advice. Because of a lack of regulation and oversight, the reliability of health care-related responses has become a topic of controversy in the medical community. To date, no study has explored the quality of AI-derived information as it relates to common foot and ankle pathologies. This study aims to assess the quality and educational benefit of ChatGPT responses to common foot and ankle-related questions. Methods ChatGPT was asked a series of 5 questions, including "What is the optimal treatment for ankle arthritis?" "How should I decide on ankle arthroplasty versus ankle arthrodesis?" "Do I need surgery for Jones fracture?" "How can I prevent Charcot arthropathy?" and "Do I need to see a doctor for my ankle sprain?" Five responses (1 per each question) were included after applying the exclusion criteria. The content was graded using DISCERN (a well-validated informational analysis tool) and AIRM (a self-designed tool for exercise evaluation). Results Health care professionals graded the ChatGPT-generated responses as bottom tier 4.5% of the time, middle tier 27.3% of the time, and top tier 68.2% of the time. Conclusion Although ChatGPT and other related AI platforms have become a popular means for medical information distribution, the educational value of the AI-generated responses related to foot and ankle pathologies was variable. With 4.5% of responses receiving a bottom-tier rating, 27.3% of responses receiving a middle-tier rating, and 68.2% of responses receiving a top-tier rating, health care professionals should be aware of the high viewership of variable-quality content easily accessible on ChatGPT. Level of Evidence Level III, cross sectional study.
Collapse
Affiliation(s)
| | - Frederic Baker Mills
- Department of Orthopaedic Surgery, Duke University Medical Center, Durham, NC, USA
| | - Mark P. Karavan
- Department of Orthopaedic Surgery, Duke University Medical Center, Durham, NC, USA
| | - Samuel B. Adams
- Department of Orthopaedic Surgery, Duke University Medical Center, Durham, NC, USA
| |
Collapse
|
27
|
Merrell LA, Fisher ND, Egol KA. Large Language Models in Orthopaedic Trauma: A Cutting-Edge Technology to Enhance the Field. J Bone Joint Surg Am 2023; 105:1383-1387. [PMID: 37402227 DOI: 10.2106/jbjs.23.00395] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 07/06/2023]
Affiliation(s)
- Lauren A Merrell
- Division of Orthopedic Trauma Surgery, Department of Orthopedic Surgery, NYU Langone Orthopedic Hospital, NYU Langone Health, New York, NY
| | | | | |
Collapse
|
28
|
Watters C, Lemanski MK. Universal skepticism of ChatGPT: a review of early literature on chat generative pre-trained transformer. Front Big Data 2023; 6:1224976. [PMID: 37680954 PMCID: PMC10482048 DOI: 10.3389/fdata.2023.1224976] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Accepted: 07/10/2023] [Indexed: 09/09/2023] Open
Abstract
ChatGPT, a new language model developed by OpenAI, has garnered significant attention in various fields since its release. This literature review provides an overview of early ChatGPT literature across multiple disciplines, exploring its applications, limitations, and ethical considerations. The review encompasses Scopus-indexed publications from November 2022 to April 2023 and includes 156 articles related to ChatGPT. The findings reveal a predominance of negative sentiment across disciplines, though subject-specific attitudes must be considered. The review highlights the implications of ChatGPT in many fields including healthcare, raising concerns about employment opportunities and ethical considerations. While ChatGPT holds promise for improved communication, further research is needed to address its capabilities and limitations. This literature review provides insights into early research on ChatGPT, informing future investigations and practical applications of chatbot technology, as well as development and usage of generative AI.
Collapse
Affiliation(s)
- Casey Watters
- Faculty of Law, Bond University, Gold Coast, QLD, Australia
| | | |
Collapse
|
29
|
Fayed AM, Mansur NSB, de Carvalho KA, Behrens A, D'Hooghe P, de Cesar Netto C. Artificial intelligence and ChatGPT in Orthopaedics and sports medicine. J Exp Orthop 2023; 10:74. [PMID: 37493985 PMCID: PMC10371934 DOI: 10.1186/s40634-023-00642-8] [Citation(s) in RCA: 18] [Impact Index Per Article: 18.0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/13/2023] [Accepted: 07/18/2023] [Indexed: 07/27/2023] Open
Abstract
Artificial intelligence (AI) is looked upon nowadays as the potential major catalyst for the fourth industrial revolution. In the last decade, AI use in Orthopaedics increased approximately tenfold. Artificial intelligence helps with tracking activities, evaluating diagnostic images, predicting injury risk, and several other uses. Chat Generated Pre-trained Transformer (ChatGPT), which is an AI-chatbot, represents an extremely controversial topic in the academic community. The aim of this review article is to simplify the concept of AI and study the extent of AI use in Orthopaedics and sports medicine literature. Additionally, the article will also evaluate the role of ChatGPT in scientific research and publications.Level of evidence: Level V, letter to review.
Collapse
Affiliation(s)
- Aly M Fayed
- Department of Orthopaedics and Rehabilitation, University of Iowa Hospitals and Clinics, Iowa City, IA, USA.
| | | | - Kepler Alencar de Carvalho
- Department of Orthopaedics and Rehabilitation, University of Iowa Hospitals and Clinics, Iowa City, IA, USA
| | - Andrew Behrens
- Department of Orthopaedics and Rehabilitation, University of Iowa Hospitals and Clinics, Iowa City, IA, USA
| | - Pieter D'Hooghe
- Aspetar Orthopedic and Sports Medicine Hospital, Doha, Qatar
| | | |
Collapse
|
30
|
Cheng K, Guo Q, He Y, Lu Y, Xie R, Li C, Wu H. Artificial Intelligence in Sports Medicine: Could GPT-4 Make Human Doctors Obsolete? Ann Biomed Eng 2023:10.1007/s10439-023-03213-1. [PMID: 37097528 DOI: 10.1007/s10439-023-03213-1] [Citation(s) in RCA: 12] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2023] [Accepted: 04/17/2023] [Indexed: 04/26/2023]
Abstract
Sports medicine, an essential branch of orthopedics, focuses on preserving, restoring, improving, and rebuilding the function of the human motor system. As a thriving interdisciplinary field, sports medicine attracts not only the interest of the orthopedic community, but also artificial intelligence (AI). In this study, our team summarized the potential applications of GPT-4 in sports medicine including diagnostic imaging, exercise prescription, medical supervision, surgery treatment, sports nutrition, and science research. In our opinion, it is impossible that GPT-4 could make sports physicians obsolete. Instead, it could become an indispensable scientific assistant for sport doctors in future.
Collapse
Affiliation(s)
- Kunming Cheng
- Department of Intensive Care Unit, The Second Affiliated Hospital of Zhengzhou University, Zhengzhou, Henan, China
| | - Qiang Guo
- Department of Orthopedics, Baodi Clinical College of Tianjin Medical University, Tianjin, China
| | - Yongbin He
- School of Sport Medicine and Rehabilitation, Beijing Sport University, Beijing, China
- University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
| | - Yanqiu Lu
- Department of Intensive Care Unit, The Second Affiliated Hospital of Zhengzhou University, Zhengzhou, Henan, China
| | - Ruijie Xie
- Department of Microsurgery, The Affiliated Nanhua Hospital, Hengyang Medical School, University of South China, Hengyang, China.
- Division of Clinical Epidemiology and Aging Research, German Cancer Research Center (DKFZ), Im Neuenheimer Feld 280, Heidelberg, Germany.
| | - Cheng Li
- Department of Orthopaedic Surgery, Beijing Jishuitan Hospital, Fourth Clinical College of Peking University, Beijing, China.
- Center for Musculoskeletal Surgery (CMSC), Charité-Universitätsmedizin Berlin, corporate member of Freie Universität Berlin, Humboldt University of Berlin, and Berlin Institute of Health, Berlin, Germany.
| | - Haiyang Wu
- Department of Graduate School, Tianjin Medical University, Tianjin, China.
- Duke Molecular Physiology Institute, Duke University School of Medicine, Durham, NC, USA.
| |
Collapse
|
31
|
He Y, Tang H, Wang D, Gu S, Ni G, Wu H. Will ChatGPT/GPT-4 be a Lighthouse to Guide Spinal Surgeons? Ann Biomed Eng 2023:10.1007/s10439-023-03206-0. [PMID: 37071280 DOI: 10.1007/s10439-023-03206-0] [Citation(s) in RCA: 22] [Impact Index Per Article: 22.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2023] [Accepted: 04/11/2023] [Indexed: 04/19/2023]
Abstract
The advent of artificial intelligence (AI), particularly ChatGPT/GPT-4, has led to advancements in various fields, including healthcare. This study explores the prospective role of ChatGPT/GPT-4 in various facets of spinal surgical practice, especially in supporting spinal surgeons during the perioperative management of endoscopic spinal surgery for patients with lumbar disc herniation. The AI-driven chatbot can facilitate communication between spinal surgeons, patients, and their relatives, streamline the collection and analysis of patient data, and contribute to the surgical planning process. Furthermore, ChatGPT/GPT-4 may enhance intraoperative support by providing real-time surgical navigation information and physiological parameter monitoring, as well as aiding in postoperative rehabilitation guidance. However, the appropriate and supervised use of ChatGPT/GPT-4 is essential, considering the potential risks associated with data security and privacy. The study concludes that ChatGPT/GPT-4 can serve as a valuable lighthouse for spinal surgeons if used correctly and responsibly.
Collapse
Affiliation(s)
- Yongbin He
- School of Sport Medicine and Rehabilitation, Beijing Sport University, Beijing, China
| | - Haifeng Tang
- School of Sport Medicine and Rehabilitation, Beijing Sport University, Beijing, China
| | - Dongxue Wang
- School of Sport Medicine and Rehabilitation, Beijing Sport University, Beijing, China
| | - Shuqin Gu
- Duke Human Vaccine Institute, Duke University Medical Center, Durham, NC, USA
| | - Guoxin Ni
- Department of Rehabilitation Medicine, The First Affiliated Hospital of Xiamen University, Xiamen, China.
| | - Haiyang Wu
- Department of Spine Surgery, Tianjin Huanhu Hospital, Graduate School of Tianjin Medical University, Tianjin, China.
- Duke Molecular Physiology Institute, Duke University School of Medicine, Durham, NC, USA.
| |
Collapse
|
32
|
Cheng K, Li Z, Li C, Xie R, Guo Q, He Y, Wu H. The Potential of GPT-4 as an AI-Powered Virtual Assistant for Surgeons Specialized in Joint Arthroplasty. Ann Biomed Eng 2023:10.1007/s10439-023-03207-z. [PMID: 37071279 DOI: 10.1007/s10439-023-03207-z] [Citation(s) in RCA: 17] [Impact Index Per Article: 17.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2023] [Accepted: 04/11/2023] [Indexed: 04/19/2023]
Abstract
The emergence of artificial intelligence (AI) offers unprecedented opportunities for joint arthroplasty surgery. Notably, on 14th March, 2023, the OpenAI company officially launched its latest version GPT-4, which once again become the focus of discussion on social media. Although more than 200 articles have reported the potential role of ChatGPT/GPT-4 in various areas, there are no studies that discussed the potential of GPT-4 as an AI-powered virtual assistant for surgeons specialized in joint arthroplasty. In this study, we summarized the five major roles of GPT-4 including scientific research, disease diagnosis, treatment options, preoperative planning, intraoperative support, and postoperative rehabilitation for arthroplasty doctors. Of note, in parallel to enjoy AI dividend, it is also necessary to pay attention to protect the data from misuse with ethical considerations in place.
Collapse
Affiliation(s)
- Kunming Cheng
- Department of Intensive Care Unit, The Second Affiliated Hospital of Zhengzhou University, Zhengzhou, Henan, China
| | - Zhiyong Li
- Department of Orthopedics, Baodi Clinical College of Tianjin Medical University, Tianjin, China
| | - Cheng Li
- Department of Orthopaedic Surgery, Beijing Jishuitan Hospital, Fourth Clinical College of Peking University, Beijing, China
| | - Ruijie Xie
- Department of Hand and Microsurgery, Hengyang Medical School, The Affiliated Nanhua Hospital, University of South China Hengyang, Hengyang, China
| | - Qiang Guo
- Department of Orthopedics, Baodi Clinical College of Tianjin Medical University, Tianjin, China
| | - Yongbin He
- School of Sport Medicine and Rehabilitation, Beijing Sport University, Beijing, China.
- University of North Carolina at Chapel Hill, Chapel Hill, NC, USA.
| | - Haiyang Wu
- Department of Graduate School, Tianjin Medical University, Tianjin, China.
- Duke Molecular Physiology Institute, Duke University School of Medicine, Durham, NC, USA.
| |
Collapse
|