1
|
Puteikis K, Mameniškienė R. Artificial intelligence: Can it help us better grasp the idea of epilepsy? An exploratory dialogue with ChatGPT and DALL·E 2. Epilepsy Behav 2024; 156:109822. [PMID: 38759427 DOI: 10.1016/j.yebeh.2024.109822] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/09/2024] [Revised: 04/29/2024] [Accepted: 05/01/2024] [Indexed: 05/19/2024]
Abstract
BACKGROUND The conceptual definition of epilepsy has been changing over decades and remains debatable. We assessed how artificial intelligence (AI) conceives epilepsy and its impact on a person's life through verbal and visual material. METHODS We asked the Chat Generative Pre-Trained Transformer (ChatGPT, OpenAI) to define epilepsy and its impact. Prompts from ChatGPT were transferred to another AI tool DALL·E 2 (Open AI) to generate visual images based on verbal input. RESULTS The ChatGPT definition on epilepsy relied on both its conceptual and practical definitions. It titled epilepsy to be "a neurological disorder characterized by recurring seizures" that has significant impact on patients' lives and is diagnosed after two or more unprovoked seizures or if there is a high risk of future seizures. ChatGPT presented nine issues - seizure-related injuries, limitations on daily activities, emotional and psychological impact, social stigma and isolation, educational and employment challenges, relationship and family dynamics, medication side effects, financial burden, and coexisting conditions - as major consequences of epilepsy. AI-generated images ranged from direct portrayals of these phenomena to abstract imagery but were mostly deprived of symbolic elements and visual metaphors. CONCLUSION We showed that AI can identify and visually interpret the burden of epilepsy from medical, societal and economical perspectives. However, the imagery created is not figurative and does not follow allegorical narratives put forward by epilepsy specialists in similar studies. The ability of AI models to lead an in-depth discussion on epilepsy remains questionable and should be explored with more advanced AI tools.
Collapse
|
2
|
AlShehri Y, McConkey M, Lodhia P. ChatGPT Provides Satisfactory but Occasionally Inaccurate Answers to Common Patient Hip Arthroscopy Questions. Arthroscopy 2024:S0749-8063(24)00452-3. [PMID: 38914299 DOI: 10.1016/j.arthro.2024.06.017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/27/2024] [Revised: 05/24/2024] [Accepted: 06/09/2024] [Indexed: 06/26/2024]
Abstract
PURPOSE To assess the ability of ChatGPT to answer common patient questions regarding hip arthroscopy, and to analyze the accuracy and appropriateness of its responses. METHODS Ten questions were selected from well-known patient education websites, and ChatGPT (version 3.5) responses to these questions were graded by 2 fellowship-trained hip preservation surgeons. Responses were analyzed, compared with the current literature, and graded from A to D (A being the highest, and D being the lowest) in a grading scale on the basis of the accuracy and completeness of the response. If the grading differed between the 2 surgeons, a consensus was reached. Inter-rater agreement was calculated. The readability of responses was also assessed using the Flesch-Kincaid Reading Ease Score (FRES) and Flesch-Kincaid Grade Level (FKGL). RESULTS Responses received the following consensus grades: A (50%, n = 5), B (30%, n = 3), C (10%, n = 1), D (10%, n = 1). Inter-rater agreement on the basis of initial individual grading was 30%. The mean FRES was 28.2 (± 9.2 standard deviation), corresponding to a college graduate level, ranging from 11.7 to 42.5. The mean FKGL was 14.4 (±1.8 standard deviation), ranging from 12.1 to 18, indicating a college student reading level. CONCLUSIONS ChatGPT can answer common patient questions regarding hip arthroscopy with satisfactory accuracy graded by 2 high-volume hip arthroscopists; however, incorrect information was identified in more than one instance. Caution must be observed when using ChatGPT for patient education related to hip arthroscopy. CLINICAL RELEVANCE Given the increasing number of hip arthroscopies being performed annually, ChatGPT has the potential to aid physicians in educating their patients about this procedure and addressing any questions they may have.
Collapse
Affiliation(s)
- Yasir AlShehri
- Department of Orthopaedics, Faculty of Medicine, The University of British Columbia, Vancouver, BC, Canada; Department of Orthopedics, College of Medicine, Imam Abdulrahman Bin Faisal University, Dammam, Saudi Arabia
| | - Mark McConkey
- Department of Orthopaedics, Faculty of Medicine, The University of British Columbia, Vancouver, BC, Canada
| | - Parth Lodhia
- Department of Orthopaedics, Faculty of Medicine, The University of British Columbia, Vancouver, BC, Canada.
| |
Collapse
|
3
|
Yu Y, Zhao Q, Zhang Y, Lin J, Wang H. Assessing the Performance of ChatGPT's Responses to Questions Related to Atopic Dermatitis. Dermatitis 2024. [PMID: 38783508 DOI: 10.1089/derm.2024.0098] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/25/2024]
Affiliation(s)
- Yan Yu
- Department of Dermatovenereology, Tianjin Medical University General Hospital/Tianjin Institute of Sexually Transmitted Disease, Tianjin, China
| | - Qian Zhao
- Department of Dermatovenereology, Tianjin Medical University General Hospital/Tianjin Institute of Sexually Transmitted Disease, Tianjin, China
| | - Yiming Zhang
- Department of Dermatovenereology, Tianjin Medical University General Hospital/Tianjin Institute of Sexually Transmitted Disease, Tianjin, China
| | - JinRu Lin
- Department of Dermatovenereology, Tianjin Medical University General Hospital/Tianjin Institute of Sexually Transmitted Disease, Tianjin, China
| | - Huiping Wang
- Department of Dermatovenereology, Tianjin Medical University General Hospital/Tianjin Institute of Sexually Transmitted Disease, Tianjin, China
| |
Collapse
|
4
|
Landais R, Sultan M, Thomas RH. The promise of AI Large Language Models for Epilepsy care. Epilepsy Behav 2024; 154:109747. [PMID: 38518673 DOI: 10.1016/j.yebeh.2024.109747] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/31/2024] [Revised: 03/08/2024] [Accepted: 03/12/2024] [Indexed: 03/24/2024]
Abstract
Artificial intelligence (AI) has been supporting our digital life for decades, but public interest in this has exploded with the recognition of large language models, such as GPT-4. We examine and evaluate the potential uses for generative AI technologies in epilepsy and neurological services. Generative AI could not only improve patient care and safety by refining communication and removing certain barriers to healthcare but may also extend to streamlining a doctor's practice through strategies such as automating paperwork. Challenges with the integration of generative AI in epilepsy services are also explored and include the risk of producing inaccurate and biased information. The impact generative AI could have on the provision of healthcare, both positive and negative, should be understood and considered carefully when deciding on the steps that need to be taken before AI is ready for use in hospitals and epilepsy services.
Collapse
Affiliation(s)
- Raphaëlle Landais
- Faculty of Medical Sciences, Newcastle University, Newcastle-Upon-Tyne NE1 7RU, United Kingdom
| | - Mustafa Sultan
- Manchester University NHS Foundation Trust, Manchester M13 9PT, United Kingdom
| | - Rhys H Thomas
- Department of Neurology, Royal Victoria Infirmary, Queen Victoria Rd, Newcastle-Upon-Tyne NE1 4LP, United Kingdom; Translational and Clinical Research Institute, Henry Wellcome Building, Framlington Place, Newcastle-Upon-Tyne NE2 4HH, United Kingdom.
| |
Collapse
|
5
|
van Diessen E, van Amerongen RA, Zijlmans M, Otte WM. Potential merits and flaws of large language models in epilepsy care: A critical review. Epilepsia 2024; 65:873-886. [PMID: 38305763 DOI: 10.1111/epi.17907] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2023] [Revised: 12/30/2023] [Accepted: 01/19/2024] [Indexed: 02/03/2024]
Abstract
The current pace of development and applications of large language models (LLMs) is unprecedented and will impact future medical care significantly. In this critical review, we provide the background to better understand these novel artificial intelligence (AI) models and how LLMs can be of future use in the daily care of people with epilepsy. Considering the importance of clinical history taking in diagnosing and monitoring epilepsy-combined with the established use of electronic health records-a great potential exists to integrate LLMs in epilepsy care. We present the current available LLM studies in epilepsy. Furthermore, we highlight and compare the most commonly used LLMs and elaborate on how these models can be applied in epilepsy. We further discuss important drawbacks and risks of LLMs, and we provide recommendations for overcoming these limitations.
Collapse
Affiliation(s)
- Eric van Diessen
- Department of Child Neurology, UMC Utrecht Brain Center, University Medical Center Utrecht and Utrecht University, Utrecht, The Netherlands
- Department of Pediatrics, Franciscus Gasthuis & Vlietland, Rotterdam, The Netherlands
| | - Ramon A van Amerongen
- Faculty of Science, Bioinformatics and Biocomplexity, Utrecht University, Utrecht, The Netherlands
| | - Maeike Zijlmans
- Department of Neurology and Neurosurgery, UMC Utrecht Brain Center, University Medical Center Utrecht and Utrecht University, Utrecht, The Netherlands
- Stichting Epilepsie Instellingen Nederland, Heemstede, The Netherlands
| | - Willem M Otte
- Department of Child Neurology, UMC Utrecht Brain Center, University Medical Center Utrecht and Utrecht University, Utrecht, The Netherlands
| |
Collapse
|
6
|
Haase I, Xiong T, Rissmann A, Knitza J, Greenfield J, Krusche M. ChatSLE: consulting ChatGPT-4 for 100 frequently asked lupus questions. THE LANCET. RHEUMATOLOGY 2024; 6:e196-e199. [PMID: 38508817 DOI: 10.1016/s2665-9913(24)00056-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/29/2024] [Revised: 02/18/2024] [Accepted: 02/19/2024] [Indexed: 03/22/2024]
Affiliation(s)
- Isabell Haase
- University Medical Center Hamburg-Eppendorf, Division of Rheumatology and Systemic Inflammatory Diseases, III Department of Medicine, 20246 Hamburg, Germany.
| | - Tingting Xiong
- University Medical Center Hamburg-Eppendorf, Division of Rheumatology and Systemic Inflammatory Diseases, III Department of Medicine, 20246 Hamburg, Germany
| | - Antonia Rissmann
- University Medical Center Hamburg-Eppendorf, Division of Rheumatology and Systemic Inflammatory Diseases, III Department of Medicine, 20246 Hamburg, Germany
| | - Johannes Knitza
- Institute for Digital Medicine, University Hospital Marburg, Philipps-University Marburg, Marburg, Germany
| | - Julia Greenfield
- Institute for Digital Medicine, University Hospital Marburg, Philipps-University Marburg, Marburg, Germany
| | - Martin Krusche
- University Medical Center Hamburg-Eppendorf, Division of Rheumatology and Systemic Inflammatory Diseases, III Department of Medicine, 20246 Hamburg, Germany
| |
Collapse
|
7
|
Huang Y, Wu R, He J, Xiang Y. Evaluating ChatGPT-4.0's data analytic proficiency in epidemiological studies: A comparative analysis with SAS, SPSS, and R. J Glob Health 2024; 14:04070. [PMID: 38547497 PMCID: PMC10978058 DOI: 10.7189/jogh.14.04070] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/02/2024] Open
Abstract
Background OpenAI's Chat Generative Pre-trained Transformer 4.0 (ChatGPT-4), an emerging artificial intelligence (AI)-based large language model (LLM), has been receiving increasing attention from the medical research community for its innovative 'Data Analyst' feature. We aimed to compare the capabilities of ChatGPT-4 against traditional biostatistical software (i.e. SAS, SPSS, R) in statistically analysing epidemiological research data. Methods We used a data set from the China Health and Nutrition Survey, comprising 9317 participants and 29 variables (e.g. gender, age, educational level, marital status, income, occupation, weekly working hours, survival status). Two researchers independently evaluated the data analysis capabilities of GPT-4's 'Data Analyst' feature against SAS, SPSS, and R across three commonly used epidemiological analysis methods: Descriptive statistics, intergroup analysis, and correlation analysis. We used an internally developed evaluation scale to assess and compare the consistency of results, analytical efficiency of coding or operations, user-friendliness, and overall performance between ChatGPT-4, SAS, SPSS, and R. Results In descriptive statistics, ChatGPT-4 showed high consistency of results, greater analytical efficiency of code or operations, and more intuitive user-friendliness compared to SAS, SPSS, and R. In intergroup comparisons and correlational analyses, despite minor discrepancies in statistical outcomes for certain analysis tasks with SAS, SPSS, and R, ChatGPT-4 maintained high analytical efficiency and exceptional user-friendliness. Thus, employing ChatGPT-4 can significantly lower the operational threshold for conducting epidemiological data analysis while maintaining consistency with traditional biostatistical software's outcome, requiring only specific, clear analysis instructions without any additional operations or code writing. Conclusions We found ChatGPT-4 to be a powerful auxiliary tool for statistical analysis in epidemiological research. However, it showed limitations in result consistency and in applying more advanced statistical methods. Therefore, we advocate for the use of ChatGPT-4 in supporting researchers with intermediate experience in data analysis. With AI technologies like LLMs advancing rapidly, their integration with data analysis platforms promises to lower operational barriers, thereby enabling researchers to dedicate greater focus to the nuanced interpretation of analysis results. This development is likely to significantly advance epidemiological and medical research.
Collapse
Affiliation(s)
- Yeen Huang
- School of Public Health and Emergency Management, Southern University of Science and Technology, Shenzhen, Guangdong, China
| | - Ruipeng Wu
- Key Laboratory for Molecular Genetic Mechanisms and Intervention Research, On High Altitude Disease of Tibet Autonomous Region, School of Medicine, Xizang Minzu University, Xianyang, Xizang, China
- Key Laboratory of High Altitude Hypoxia Environment and Life Health, School of Medicine, Xizang Minzu University, Xianyang, Xizang, China
- Key Laboratory of Environmental Medicine and Engineering of Ministry of Education, Department of Nutrition and Food Hygiene, School of Public Health, Southeast University, Nanjing, Jiangsu, China
| | - Juntao He
- Physical and Chemical Testing Institute, Shenzhen Prevention and Treatment Center for Occupational Diseases, Shenzhen, Guangdong, China
| | - Yingping Xiang
- Occupational Hazard Assessment Institute, Shenzhen Prevention and Treatment Center for Occupational Diseases, Shenzhen, Guangdong, China
| |
Collapse
|
8
|
Daungsupawong H, Wiwanitkit V. ChatGPT's responses to questions related to epilepsy. Seizure 2024; 114:105. [PMID: 38118283 DOI: 10.1016/j.seizure.2023.12.004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2023] [Accepted: 12/04/2023] [Indexed: 12/22/2023] Open
Abstract
This is a correspondence on published article on "ChatGPT's responses to questions related to epilepsy".
Collapse
Affiliation(s)
| | - Viroj Wiwanitkit
- Department of Research Analytics, Saveetha Dental College and Hospitals, Saveetha Institute of Medical and Technical Sciences Saveetha University India
| |
Collapse
|