1
|
B.T B, Chen JM. Performance Assessment of ChatGPT versus Bard in Detecting Alzheimer's Dementia. Diagnostics (Basel) 2024; 14:817. [PMID: 38667463 PMCID: PMC11048951 DOI: 10.3390/diagnostics14080817] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2024] [Revised: 04/08/2024] [Accepted: 04/10/2024] [Indexed: 04/28/2024] Open
Abstract
Large language models (LLMs) find increasing applications in many fields. Here, three LLM chatbots (ChatGPT-3.5, ChatGPT-4, and Bard) are assessed in their current form, as publicly available, for their ability to recognize Alzheimer's dementia (AD) and Cognitively Normal (CN) individuals using textual input derived from spontaneous speech recordings. A zero-shot learning approach is used at two levels of independent queries, with the second query (chain-of-thought prompting) eliciting more detailed information than the first. Each LLM chatbot's performance is evaluated on the prediction generated in terms of accuracy, sensitivity, specificity, precision, and F1 score. LLM chatbots generated a three-class outcome ("AD", "CN", or "Unsure"). When positively identifying AD, Bard produced the highest true-positives (89% recall) and highest F1 score (71%), but tended to misidentify CN as AD, with high confidence (low "Unsure" rates); for positively identifying CN, GPT-4 resulted in the highest true-negatives at 56% and highest F1 score (62%), adopting a diplomatic stance (moderate "Unsure" rates). Overall, the three LLM chatbots can identify AD vs. CN, surpassing chance-levels, but do not currently satisfy the requirements for clinical application.
Collapse
Affiliation(s)
- Balamurali B.T
- Science, Mathematics & Technology (SMT), Singapore University of Technology & Design, 8 Somapah Rd, Singapore 487372, Singapore
| | - Jer-Ming Chen
- Science, Mathematics & Technology (SMT), Singapore University of Technology & Design, 8 Somapah Rd, Singapore 487372, Singapore
| |
Collapse
|
2
|
LUZ SATURNINO, HAIDER FASIH, FROMM DAVIDA, LAZAROU IOULIETTA, KOMPATSIARIS IOANNIS, MACWHINNEY BRIAN. An Overview of the ADReSS-M Signal Processing Grand Challenge on Multilingual Alzheimer's Dementia Recognition Through Spontaneous Speech. IEEE OPEN JOURNAL OF SIGNAL PROCESSING 2024; 5:738-749. [PMID: 38957540 PMCID: PMC11218814 DOI: 10.1109/ojsp.2024.3378595] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 07/04/2024]
Abstract
The ADReSS-M Signal Processing Grand Challenge was held at the 2023 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2023. The challenge targeted difficult automatic prediction problems of great societal and medical relevance, namely, the detection of Alzheimer's Dementia (AD) and the estimation of cognitive test scoress. Participants were invited to create models for the assessment of cognitive function based on spontaneous speech data. Most of these models employed signal processing and machine learning methods. The ADReSS-M challenge was designed to assess the extent to which predictive models built based on speech in one language generalise to another language. The language data compiled and made available for ADReSS-M comprised English, for model training, and Greek, for model testing and validation. To the best of our knowledge no previous shared research task investigated acoustic features of the speech signal or linguistic characteristics in the context of multilingual AD detection. This paper describes the context of the ADReSS-M challenge, its data sets, its predictive tasks, the evaluation methodology we employed, our baseline models and results, and the top five submissions. The paper concludes with a summary discussion of the ADReSS-M results, and our critical assessment of the future outlook in this field.
Collapse
Affiliation(s)
- SATURNINO LUZ
- Usher Institute, Edinburgh Medical School, The University of Edinburgh, EH16 4UX Edinburgh, U.K
| | - FASIH HAIDER
- School of Engineering, The University of Edinburgh, EH9 3JW Edinburgh, U.K
| | - DAVIDA FROMM
- Department of Psychology, Carnegie Mellon University, Pittsburgh 15213, PA USA
| | - IOULIETTA LAZAROU
- Information Technologies Institute, CERTH, Thessaloniki, Thermi-Thessaloniki 57001, Greece
| | - IOANNIS KOMPATSIARIS
- Information Technologies Institute, CERTH, Thessaloniki, Thermi-Thessaloniki 57001, Greece
| | - BRIAN MACWHINNEY
- Department of Psychology, Carnegie Mellon University, Pittsburgh 15213, PA USA
| |
Collapse
|
3
|
Liu N, Wang L. An approach for assisting diagnosis of Alzheimer's disease based on natural language processing. Front Aging Neurosci 2023; 15:1281726. [PMID: 38035270 PMCID: PMC10687444 DOI: 10.3389/fnagi.2023.1281726] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2023] [Accepted: 10/17/2023] [Indexed: 12/02/2023] Open
Abstract
Introduction Alzheimer's Disease (AD) is a common dementia which affects linguistic function, memory, cognitive and visual spatial ability of the patients. Language is proved to have the relationship with AD, so the time that AD can be diagnosed in a doctor's office is coming. Methods In this study, the Pitt datasets are used to detect AD which is balanced in gender and age. First bidirectional Encoder Representation from Transformers (Bert) pretrained model is used to acquire the word vector. Then two channels are constructed in the feature extraction layer, which is, convolutional neural networks (CNN) and long and short time memory (LSTM) model to extract local features and global features respectively. The local features and global features are concatenated to generate feature vectors containing rich semantics, which are sent to softmax classifier for classification. Results Finally, we obtain a best accuracy of 89.3% which is comparative compared to other studies. In the meanwhile, we do the comparative experiments with TextCNN and LSTM model respectively, the combined model manifests best and TextCNN takes the second place. Discussion The performance illustrates the feasibility to predict AD effectively by using acoustic and linguistic datasets.
Collapse
Affiliation(s)
- Ning Liu
- School of Science/School of Big Data Science, Zhejiang University of Science and Technology, Zhejiang, China
| | - Lingxing Wang
- Department of Neurology, Second Affiliated Hospital of Fujian Medical University, Quanzhou, Fujian, China
| |
Collapse
|
4
|
Parsapoor M. AI-based assessments of speech and language impairments in dementia. Alzheimers Dement 2023; 19:4675-4687. [PMID: 37578167 DOI: 10.1002/alz.13395] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2022] [Revised: 06/03/2023] [Accepted: 06/05/2023] [Indexed: 08/15/2023]
Abstract
Recent advancements in the artificial intelligence (AI) domain have revolutionized the early detection of cognitive impairments associated with dementia. This has motivated clinicians to use AI-powered dementia detection systems, particularly systems developed based on individuals' and patients' speech and language, for a quick and accurate identification of patients with dementia. This paper reviews articles about developing assessment tools using machine learning and deep learning algorithms trained by vocal and textual datasets.
Collapse
Affiliation(s)
- Mahboobeh Parsapoor
- Centre de Recherche Informatique de Montréal: CRIM, Montreal, Quebec, Canada
| |
Collapse
|
5
|
Qi X, Zhou Q, Dong J, Bao W. Noninvasive automatic detection of Alzheimer's disease from spontaneous speech: a review. Front Aging Neurosci 2023; 15:1224723. [PMID: 37693647 PMCID: PMC10484224 DOI: 10.3389/fnagi.2023.1224723] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Accepted: 08/04/2023] [Indexed: 09/12/2023] Open
Abstract
Alzheimer's disease (AD) is considered as one of the leading causes of death among people over the age of 70 that is characterized by memory degradation and language impairment. Due to language dysfunction observed in individuals with AD patients, the speech-based methods offer non-invasive, convenient, and cost-effective solutions for the automatic detection of AD. This paper systematically reviews the technologies to detect the onset of AD from spontaneous speech, including data collection, feature extraction and classification. First the paper formulates the task of automatic detection of AD and describes the process of data collection. Then, feature extractors from speech data and transcripts are reviewed, which mainly contains acoustic features from speech and linguistic features from text. Especially, general handcrafted features and deep embedding features are organized from different modalities. Additionally, this paper summarizes optimization strategies for AD detection systems. Finally, the paper addresses challenges related to data size, model explainability, reliability and multimodality fusion, and discusses potential research directions based on these challenges.
Collapse
Affiliation(s)
- Xiaoke Qi
- School of Information Management for Law, China University of Political Science and Law, Beijing, China
| | | | - Jian Dong
- Information Technology Research Center, China Electronics Standardization Institute, Beijing, China
| | - Wei Bao
- Information Technology Research Center, China Electronics Standardization Institute, Beijing, China
| |
Collapse
|
6
|
Liu N, Yuan Z, Chen Y, Liu C, Wang L. Learning implicit sentiments in Alzheimer's disease recognition with contextual attention features. Front Aging Neurosci 2023; 15:1122799. [PMID: 37266402 PMCID: PMC10231228 DOI: 10.3389/fnagi.2023.1122799] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2022] [Accepted: 04/05/2023] [Indexed: 06/03/2023] Open
Abstract
Background Alzheimer's disease (AD) is difficult to diagnose on the basis of language because of the implicit emotion of transcripts, which is defined as a supervised fuzzy implicit emotion classification at the document level. Recent neural network-based approaches have not paid attention to the implicit sentiments entailed in AD transcripts. Method A two-level attention mechanism is proposed to detect deep semantic information toward words and sentences, which enables it to attend to more words and fewer sentences differentially when constructing document representation. Specifically, a document vector was built by progressively aggregating important words into sentence vectors and important sentences into document vectors. Results Experimental results showed that our method achieved the best accuracy of 91.6% on annotated public Pitt corpora, which validates its effectiveness in learning implicit sentiment representation for our model. Conclusion The proposed model can qualitatively select informative words and sentences using attention layers, and this method also provides good inspiration for AD diagnosis based on implicit sentiment transcripts.
Collapse
Affiliation(s)
- Ning Liu
- School of Science/School of Big Data Science, Zhejiang University of Science and Technology, Hangzhou, China
| | - Zhenming Yuan
- School of Information Science and Technology, Hangzhou Normal University, Hangzhou, Zhejiang, China
| | - Yan Chen
- International Unresponsive Wakefulness Syndrome and Consciousness Science Institute, Hangzhou Normal University, Hangzhou, China
| | - Chuan Liu
- School of Mathematics and Computer Science, Quanzhou Normal University, Quanzhou, Fujian, China
| | - Lingxing Wang
- Department of Neurology, Second Affiliated Hospital of Fujian Medical University, Quanzhou, Fujian, China
| |
Collapse
|
7
|
Alfalahi H, Dias SB, Khandoker AH, Chaudhuri KR, Hadjileontiadis LJ. A scoping review of neurodegenerative manifestations in explainable digital phenotyping. NPJ Parkinsons Dis 2023; 9:49. [PMID: 36997573 PMCID: PMC10063633 DOI: 10.1038/s41531-023-00494-0] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/20/2022] [Accepted: 03/16/2023] [Indexed: 04/03/2023] Open
Abstract
Neurologists nowadays no longer view neurodegenerative diseases, like Parkinson's and Alzheimer's disease, as single entities, but rather as a spectrum of multifaceted symptoms with heterogeneous progression courses and treatment responses. The definition of the naturalistic behavioral repertoire of early neurodegenerative manifestations is still elusive, impeding early diagnosis and intervention. Central to this view is the role of artificial intelligence (AI) in reinforcing the depth of phenotypic information, thereby supporting the paradigm shift to precision medicine and personalized healthcare. This suggestion advocates the definition of disease subtypes in a new biomarker-supported nosology framework, yet without empirical consensus on standardization, reliability and interpretability. Although the well-defined neurodegenerative processes, linked to a triad of motor and non-motor preclinical symptoms, are detected by clinical intuition, we undertake an unbiased data-driven approach to identify different patterns of neuropathology distribution based on the naturalistic behavior data inherent to populations in-the-wild. We appraise the role of remote technologies in the definition of digital phenotyping specific to brain-, body- and social-level neurodegenerative subtle symptoms, emphasizing inter- and intra-patient variability powered by deep learning. As such, the present review endeavors to exploit digital technologies and AI to create disease-specific phenotypic explanations, facilitating the understanding of neurodegenerative diseases as "bio-psycho-social" conditions. Not only does this translational effort within explainable digital phenotyping foster the understanding of disease-induced traits, but it also enhances diagnostic and, eventually, treatment personalization.
Collapse
Affiliation(s)
- Hessa Alfalahi
- Department of Biomedical Engineering, Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates.
- Healthcare Engineering Innovation Center (HEIC), Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates.
| | - Sofia B Dias
- Department of Biomedical Engineering, Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates
- Healthcare Engineering Innovation Center (HEIC), Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates
- CIPER, Faculdade de Motricidade Humana, University of Lisbon, Lisbon, Portugal
| | - Ahsan H Khandoker
- Department of Biomedical Engineering, Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates
- Healthcare Engineering Innovation Center (HEIC), Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates
| | - Kallol Ray Chaudhuri
- Parkinson Foundation, International Center of Excellence, King's College London, Denmark Hills, London, UK
- Department of Basic and Clinical Neurosciences, Institute of Psychiatry, Psychology and Neuroscience, King's College London, De Crespigny Park, London, UK
| | - Leontios J Hadjileontiadis
- Department of Biomedical Engineering, Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates
- Healthcare Engineering Innovation Center (HEIC), Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates
- Department of Electrical and Computer Engineering, Aristotle University of Thessaloniki, Thessaloniki, Greece
| |
Collapse
|
8
|
Parsapoor (Parsa) M(M, Alam MR, Mihailidis A. Performance of machine learning algorithms for dementia assessment: impacts of language tasks, recording media, and modalities. BMC Med Inform Decis Mak 2023; 23:45. [PMID: 36869377 PMCID: PMC9985301 DOI: 10.1186/s12911-023-02122-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2021] [Accepted: 01/23/2023] [Indexed: 03/05/2023] Open
Abstract
OBJECTIVES Automatic speech and language assessment methods (SLAMs) can help clinicians assess speech and language impairments associated with dementia in older adults. The basis of any automatic SLAMs is a machine learning (ML) classifier that is trained on participants' speech and language. However, language tasks, recording media, and modalities impact the performance of ML classifiers. Thus, this research has focused on evaluating the effects of the above-mentioned factors on the performance of ML classifiers that can be used for dementia assessment. METHODOLOGY Our methodology includes the following steps: (1) Collecting speech and language datasets from patients and healthy controls; (2) Using feature engineering methods which include feature extraction methods to extract linguistic and acoustic features and feature selection methods to select most informative features; (3) Training different ML classifiers; and (4) Evaluating the performance of ML classifiers to investigate the impacts of language tasks, recording media, and modalities on dementia assessment. RESULTS Our results show that (1) the ML classifiers trained with the picture description language task perform better than the classifiers trained with the story recall language task; (2) the data obtained from phone-based recordings improves the performance of ML classifiers compared to data obtained from web-based recordings; and (3) the ML classifiers trained with acoustic features perform better than the classifiers trained with linguistic features. CONCLUSION This research demonstrates that we can improve the performance of automatic SLAMs as dementia assessment methods if we: (1) Use the picture description task to obtain participants' speech; (2) Collect participants' voices via phone-based recordings; and (3) Train ML classifiers using only acoustic features. Our proposed methodology will help future researchers to investigate the impacts of different factors on the performance of ML classifiers for assessing dementia.
Collapse
Affiliation(s)
| | - Muhammad Raisul Alam
- grid.17063.330000 0001 2157 2938Department of Computer Science, University of Toronto, Toronto, Canada
- grid.17063.330000 0001 2157 2938Department Occupational Science and Occupational Therapy, University of Toronto, Toronto, Canada
- grid.494618.6Vector Institute, Toronto, Canada
| | - Alex Mihailidis
- grid.17063.330000 0001 2157 2938Department Occupational Science and Occupational Therapy, University of Toronto, Toronto, Canada
- grid.17063.330000 0001 2157 2938Institute of Biomedical Engineering, University of Toronto, Toronto, Canada
| |
Collapse
|
9
|
Detecting dementia from speech and transcripts using transformers. COMPUT SPEECH LANG 2023. [DOI: 10.1016/j.csl.2023.101485] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/15/2023]
|
10
|
Ilias L, Askounis D. Explainable Identification of Dementia from Transcripts using Transformer Networks. IEEE J Biomed Health Inform 2022; 26:4153-4164. [PMID: 35511841 DOI: 10.1109/jbhi.2022.3172479] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Alzheimers disease (AD) is the main cause of dementia which is accompanied by loss of memory and may lead to severe consequences in peoples everyday life if not diagnosed on time. Very few works have exploited transformer-based networks and despite the high accuracy achieved, little work has been done in terms of model interpretability. In addition, although Mini-Mental State Exam (MMSE) scores are inextricably linked with the identification of dementia, research works face the task of dementia identification and the task of the prediction of MMSE scores as two separate tasks. In order to address these limitations, we employ several transformer-based models, with BERT achieving the highest accuracy accounting for 87.50%. Concurrently, we propose an interpretable method to detect AD patients based on siamese networks reaching accuracy up to 83.75%. Next, we introduce two multi-task learning models, where the main task refers to the identification of dementia (binary classification), while the auxiliary one corresponds to the identification of the severity of dementia (multiclass classification). Our model obtains accuracy equal to 86.25% on the detection of AD patients in the multi-task learning setting. Finally, we present some new methods to identify the linguistic patterns used by AD patients and non-AD ones, including text statistics, vocabulary uniqueness, word usage, correlations via a detailed linguistic analysis, and explainability techniques (LIME). Findings indicate significant differences in language between AD and non-AD patients.
Collapse
|
11
|
Ilias L, Askounis D. Multimodal Deep Learning Models for Detecting Dementia From Speech and Transcripts. Front Aging Neurosci 2022; 14:830943. [PMID: 35370608 PMCID: PMC8969102 DOI: 10.3389/fnagi.2022.830943] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2021] [Accepted: 02/17/2022] [Indexed: 11/13/2022] Open
Abstract
Alzheimer's dementia (AD) entails negative psychological, social, and economic consequences not only for the patients but also for their families, relatives, and society in general. Despite the significance of this phenomenon and the importance for an early diagnosis, there are still limitations. Specifically, the main limitation is pertinent to the way the modalities of speech and transcripts are combined in a single neural network. Existing research works add/concatenate the image and text representations, employ majority voting approaches or average the predictions after training many textual and speech models separately. To address these limitations, in this article we present some new methods to detect AD patients and predict the Mini-Mental State Examination (MMSE) scores in an end-to-end trainable manner consisting of a combination of BERT, Vision Transformer, Co-Attention, Multimodal Shifting Gate, and a variant of the self-attention mechanism. Specifically, we convert audio to Log-Mel spectrograms, their delta, and delta-delta (acceleration values). First, we pass each transcript and image through a BERT model and Vision Transformer, respectively, adding a co-attention layer at the top, which generates image and word attention simultaneously. Secondly, we propose an architecture, which integrates multimodal information to a BERT model via a Multimodal Shifting Gate. Finally, we introduce an approach to capture both the inter- and intra-modal interactions by concatenating the textual and visual representations and utilizing a self-attention mechanism, which includes a gate model. Experiments conducted on the ADReSS Challenge dataset indicate that our introduced models demonstrate valuable advantages over existing research initiatives achieving competitive results in both the AD classification and MMSE regression task. Specifically, our best performing model attains an accuracy of 90.00% and a Root Mean Squared Error (RMSE) of 3.61 in the AD classification task and MMSE regression task, respectively, achieving a new state-of-the-art performance in the MMSE regression task.
Collapse
Affiliation(s)
- Loukas Ilias
- Decision Support Systems Laboratory, School of Electrical and Computer Engineering, National Technical University of Athens, Athens, Greece
| | | |
Collapse
|