1
|
Botchu B, Iyengar KP, Botchu R. Can ChatGPT empower people with dyslexia? Disabil Rehabil Assist Technol 2024; 19:2131-2132. [PMID: 37697967 DOI: 10.1080/17483107.2023.2256805] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2023] [Accepted: 09/05/2023] [Indexed: 09/13/2023]
Affiliation(s)
| | - Karthikeyan P Iyengar
- Department of Orthopaedics, Southport and Ormskirk Hospital NHS Trust, Southport, UK
| | - Rajesh Botchu
- Department of Musculoskeletal Radiology, Royal Orthopedic Hospital, Birmingham, UK
| |
Collapse
|
2
|
Costa ICP, do Nascimento MC, Treviso P, Chini LT, Roza BDA, Barbosa SDFF, Mendes KDS. Using the Chat Generative Pre-trained Transformer in academic writing in health: a scoping review. Rev Lat Am Enfermagem 2024; 32:e4194. [PMID: 38922265 PMCID: PMC11182606 DOI: 10.1590/1518-8345.7133.4194] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2023] [Accepted: 02/04/2024] [Indexed: 06/27/2024] Open
Abstract
OBJECTIVE to map the scientific literature regarding the use of the Chat Generative Pre-trained Transformer, ChatGPT, in academic writing in health. METHOD this was a scoping review, following the JBI methodology. Conventional databases and gray literature were included. The selection of studies was applied after removing duplicates and individual and paired evaluation. Data were extracted based on an elaborate script, and presented in a descriptive, tabular and graphical format. RESULTS the analysis of the 49 selected articles revealed that ChatGPT is a versatile tool, contributing to scientific production, description of medical procedures and preparation of summaries aligned with the standards of scientific journals. Its application has been shown to improve the clarity of writing and benefits areas such as innovation and automation. Risks were also observed, such as the possibility of lack of originality and ethical issues. Future perspectives highlight the need for adequate regulation, agile adaptation and the search for an ethical balance in incorporating ChatGPT into academic writing. CONCLUSION ChatGPT presents transformative potential in academic writing in health. However, its adoption requires rigorous human supervision, solid regulation, and transparent guidelines to ensure its responsible and beneficial use by the scientific community.
Collapse
Affiliation(s)
| | | | - Patrícia Treviso
- Universidade do Vale do Rio dos Sinos, Escola de Saúde, São Leopoldo, RS, Brazil
| | | | | | | | - Karina Dal Sasso Mendes
- Universidade de São Paulo, Escola de Enfermagem de Ribeirão Preto, PAHO/WHO Collaborating Centre for Nursing Research Development, Ribeirão Preto, SP, Brazil
| |
Collapse
|
3
|
Jenko N, Ariyaratne S, Jeys L, Evans S, Iyengar KP, Botchu R. An evaluation of AI generated literature reviews in musculoskeletal radiology. Surgeon 2024; 22:194-197. [PMID: 38218659 DOI: 10.1016/j.surge.2023.12.005] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/11/2023] [Revised: 12/20/2023] [Accepted: 12/27/2023] [Indexed: 01/15/2024]
Abstract
PURPOSE The use of artificial intelligence (AI) tools to aid in summarizing information in medicine and research has recently garnered a huge amount of interest. While tools such as ChatGPT produce convincing and naturally sounding output, the answers are sometimes incorrect. Some of these drawbacks, it is hoped, can be avoided by using programmes trained for a more specific scope. In this study we compared the performance of a new AI tool (the-literature.com) to the latest version OpenAI's ChatGPT (GPT-4) in summarizing topics that the authors have significantly contributed to. METHODS The AI tools were asked to produce a literature review on 7 topics. These were selected based on the research topics that the authors were intimately familiar with and have contributed to through their own publications. The output produced by the AI tools were graded on a 1-5 Likert scale for accuracy, comprehensiveness, and relevance by two fellowship trained consultant radiologists. RESULTS The-literature.com produced 3 excellent summaries, 3 very poor summaries not relevant to the prompt, and one summary, which was relevant but did not include all relevant papers. All of the summaries produced by GPT-4 were relevant, but fewer relevant papers were identified. The average Likert rating was for the-literature was 2.88 and 3.86 for GPT-4. There was good agreement between the ratings of both radiologists (ICC = 0.883). CONCLUSION Summaries produced by AI in its current state require careful human validation. GPT-4 on average provides higher quality summaries. Neither tool can reliably identify all relevant publications.
Collapse
Affiliation(s)
- N Jenko
- Radiology, Royal Orthopaedic Hospital NHS Foundation Trust, Birmingham, UK.
| | - S Ariyaratne
- Radiology, Royal Orthopaedic Hospital NHS Foundation Trust, Birmingham, UK
| | - L Jeys
- Orthopaedic Surgery, Royal Orthopaedic Hospital NHS Foundation Trust, Birmingham, UK
| | - S Evans
- Orthopaedic Surgery, Royal Orthopaedic Hospital NHS Foundation Trust, Birmingham, UK
| | - K P Iyengar
- Orthopaedic Surgery, Mersey and West Lancashire Teaching Hospitals NHS Trust, Southport, UK
| | - R Botchu
- Radiology, Royal Orthopaedic Hospital NHS Foundation Trust, Birmingham, UK
| |
Collapse
|
4
|
Ariyaratne S, Jenko N, Mark Davies A, Iyengar KP, Botchu R. Could ChatGPT Pass the UK Radiology Fellowship Examinations? Acad Radiol 2024; 31:2178-2182. [PMID: 38160089 DOI: 10.1016/j.acra.2023.11.026] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2023] [Revised: 11/12/2023] [Accepted: 11/18/2023] [Indexed: 01/03/2024]
Abstract
RATIONALE AND OBJECTIVES Chat Generative Pre-trained Transformer (ChatGPT) is an artificial intelligence (AI) tool which utilises machine learning to generate original text resembling human language. AI models have recently demonstrated remarkable ability at analysing and solving problems, including passing professional examinations. We investigate the performance of ChatGPT on some of the UK radiology fellowship equivalent examination questions. METHODS ChatGPT was asked to answer questions from question banks resembling the Fellowship of the Royal College of Radiologists (FRCR) examination. The entire physics part 1 question bank (203 5-part true/false questions) was answered by the GPT-4 model and answers recorded. 240 single best answer questions (SBAs) (representing the true length of the FRCR 2A examination) were answered by both GPT-3.5 and GPT-4 models. RESULTS ChatGPT 4 answered 74.8% of part 1 true/false statements correctly. The spring 2023 passing mark of the part 1 examination was 75.5% and ChatGPT thus narrowly failed. In the 2A examination, ChatGPT 3.5 answered 50.8% SBAs correctly, while GPT-4 answered 74.2% correctly. The winter 2022 2A pass mark was 63.3% and thus GPT-4 clearly passed. CONCLUSION AI models such as ChatGPT are able to answer the majority of questions in an FRCR style examination. It is reasonable to assume that further developments in AI will be more likely to succeed in comprehending and solving questions related to medicine, specifically clinical radiology. ADVANCES IN KNOWLEDGE Our findings outline the unprecedented capabilities of AI, adding to the current relatively small body of literature on the subject, which in turn can play a role medical training, evaluation and practice. This can undoubtedly have implications for radiology.
Collapse
Affiliation(s)
- Sisith Ariyaratne
- Department of Musculoskeletal Radiology, The Royal Orthopaedic Hospital NHS Foundation Trust, Northfield, Birmingham, UK (S.A., N.J., A.M.D., R.B.).
| | - Nathan Jenko
- Department of Musculoskeletal Radiology, The Royal Orthopaedic Hospital NHS Foundation Trust, Northfield, Birmingham, UK (S.A., N.J., A.M.D., R.B.)
| | - A Mark Davies
- Department of Musculoskeletal Radiology, The Royal Orthopaedic Hospital NHS Foundation Trust, Northfield, Birmingham, UK (S.A., N.J., A.M.D., R.B.)
| | - Karthikeyan P Iyengar
- Department of Trauma & Orthopaedics, Southport & Ormskirk Hospitals, Mersey and West Lancashire Teaching NHS Trust, Southport, UK (K.P.I.)
| | - Rajesh Botchu
- Department of Musculoskeletal Radiology, The Royal Orthopaedic Hospital NHS Foundation Trust, Northfield, Birmingham, UK (S.A., N.J., A.M.D., R.B.)
| |
Collapse
|
5
|
Saad A, Jenko N, Ariyaratne S, Birch N, Iyengar KP, Davies AM, Vaishya R, Botchu R. Exploring the potential of ChatGPT in the peer review process: An observational study. Diabetes Metab Syndr 2024; 18:102946. [PMID: 38330745 DOI: 10.1016/j.dsx.2024.102946] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/21/2023] [Revised: 01/09/2024] [Accepted: 01/10/2024] [Indexed: 02/10/2024]
Abstract
BACKGROUND Peer review is the established method for evaluating the quality and validity of research manuscripts in scholarly publishing. However, scientific peer review faces challenges as the volume of submitted research has steadily increased in recent years. Time constraints and peer review quality assurance can place burdens on reviewers, potentially discouraging their participation. Some artificial intelligence (AI) tools might assist in relieving these pressures. This study explores the efficiency and effectiveness of one of the artificial intelligence (AI) chatbots, ChatGPT (Generative Pre-trained Transformer), in the peer review process. METHODS Twenty-one peer-reviewed research articles were anonymised to ensure unbiased evaluation. Each article was reviewed by two humans and by versions 3.5 and 4.0 of ChatGPT. The AI was instructed to provide three positive and three negative comments on the articles and recommend whether they should be accepted or rejected. The human and AI results were compared using a 5-point Likert scale to determine the level of agreement. The correlation between ChatGPT responses and the acceptance or rejection of the papers was also examined. RESULTS Subjective review similarity between human reviewers and ChatGPT showed a mean score of 3.6/5 for ChatGPT 3.5 and 3.76/5 for ChatGPT 4.0. The correlation between human and AI review scores was statistically significant for ChatGPT 3.5, but not for ChatGPT 4.0. CONCLUSION ChatGPT can complement human scientific peer review, enhancing efficiency and promptness in the editorial process. However, a fully automated AI review process is currently not advisable, and ChatGPT's role should be regarded as highly constrained for the present and near future.
Collapse
Affiliation(s)
- Ahmed Saad
- Department of Orthopedics, Royal Orthopaedic Hospital, Birmingham, B31 2AP, UK.
| | - Nathan Jenko
- Department of Musculoskeletal Radiology, Royal Orthopaedic Hospital, Birmingham, B31 2AP, UK.
| | - Sisith Ariyaratne
- Department of Musculoskeletal Radiology, Royal Orthopaedic Hospital, Birmingham, B31 2AP, UK.
| | - Nick Birch
- East Midlands Spine, Bragborough Hall Health & Wellbeing Centre, Welton Road, Braunston, Daventry, Northants, NN117JG, UK.
| | - Karthikeyan P Iyengar
- Department of Orthopedics, Mersey and West Lancashire Teaching Hospitals NHS Trust, Southport, PR8 6PN, UK.
| | - Arthur Mark Davies
- Department of Musculoskeletal Radiology, Royal Orthopaedic Hospital, Birmingham, B31 2AP, UK.
| | - Raju Vaishya
- Department of Orthopedics, Indraprastha Apollo Hospital, Mathura Rd, New Delhi, 110076, India.
| | - Rajesh Botchu
- Department of Musculoskeletal Radiology, Royal Orthopaedic Hospital, Birmingham, B31 2AP, UK.
| |
Collapse
|
6
|
Saad A, Iyengar KP, Kurisunkal V, Botchu R. Assessing ChatGPT's ability to pass the FRCS orthopaedic part A exam: A critical analysis. Surgeon 2023; 21:263-266. [PMID: 37517980 DOI: 10.1016/j.surge.2023.07.001] [Citation(s) in RCA: 9] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/01/2023] [Accepted: 07/03/2023] [Indexed: 08/01/2023]
Abstract
AI technology has made significant advancements in recent years, with the notable development of ChatGPT in November 2022. Users have observed evidence of deductive reasoning, logical thinking, and coherent thought in ChatGPT's responses. This study aimed to determine if ChatGPT has the capability to pass the Orthopaedic Fellow of the Royal College of Surgeons (FRCS Orth) Part A exam. METHODS To assess ChatGPT4's ability to pass the Orthopaedic FRCS Orth Part A exam, a study was conducted using 240 mock FRCS Orth Part A questions. The study evaluated the accuracy of ChatGPT's answers and the response time for each question. Descriptive statistics were employed to analyse the chatbot's performance. RESULTS The evaluation revealed that ChatGPT4 achieved an overall score of 67.5% on Part A of the exam. However, ChatGPT4 did not meet the overall pass mark required for the FRCS Orth Part A exam. CONCLUSION This study demonstrates that ChatGPT was unable to pass the FRCS Orthopaedic examination. Several factors contributed to this outcome, including the lack of critical or high-order thinking abilities, limited clinical expertise, and the inability to meet the rigorous requirements of the exam.
Collapse
Affiliation(s)
- Ahmed Saad
- Department of Orthopedics, Royal Orthopedic Hospital, Birmingham, UK.
| | - Karthikeyan P Iyengar
- Department of Orthopedics, Southport and Ormskirk Hospital NHS Trust, Southport, UK.
| | - Vineet Kurisunkal
- Department of Orthopedic Oncology, Royal Orthopedic Hospital, Birmingham, UK.
| | - Rajesh Botchu
- Department of Musculoskeletal Radiology, Royal Orthopedic Hospital, Birmingham, UK.
| |
Collapse
|