1
|
Park Y, Park S, Lee M. Effectiveness of artificial intelligence in detecting and managing depressive disorders: Systematic review. J Affect Disord 2024; 361:445-456. [PMID: 38889858 DOI: 10.1016/j.jad.2024.06.035] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/27/2023] [Revised: 11/27/2023] [Accepted: 06/14/2024] [Indexed: 06/20/2024]
Abstract
OBJECTIVES This study underscores the importance of exploring AI's creative applications in treating depressive disorders to revolutionize mental health care. Through innovative integration of AI technologies, the research confirms their positive effects on preventing, diagnosing, and treating depression. The systematic review establishes an evidence base for AI in depression management, offering directions for effective interventions. METHODS This systematic literature review investigates the effectiveness of AI in depression management by analyzing studies from January 1, 2017, to May 31, 2022. Utilizing search engines like IEEE Xplore, PubMed, and Web of Science, the review focused on keywords such as Depression/Mental Health, Machine Learning/Artificial Intelligence, and Prediction/Diagnosis. The analysis of 95 documents involved classification based on use, data type, and algorithm type. RESULTS The study revealed that AI in depression management excelled in accuracy, particularly in monitoring and prediction. Biomarker-derived data demonstrated the highest accuracy, with the CNN algorithm proving most effective. The findings affirm the therapeutic benefits of AI, including treatment, detection, and disease prediction, highlighting its potential in analyzing monitored data for depression management. LIMITATIONS This study exclusively examined the application of AI in individuals with depressive disorders. Interpretation should be cautious due to the limited scope of subjects to this specific population. CONCLUSIONS To introduce digital healthcare and therapies for ongoing depression management, it's crucial to present empirical evidence on the medical fee payment system, safety, and efficacy. These findings support enhanced medical accessibility through digital healthcare, offering personalized disease management for patients seeking non-face-to-face treatment.
Collapse
Affiliation(s)
- Yoonseo Park
- Department of Convergence Healthcare Medicine, Ajou University, Suwon, South Korea
| | - Sewon Park
- Department of Medical Science, Ajou University School of Medicine, Suwon, South Korea
| | - Munjae Lee
- Department of Medical Science, Ajou University School of Medicine, Suwon, South Korea.
| |
Collapse
|
2
|
Simon GE, Cruz M, Boggs JM, Beck A, Shortreed SM, Coley RY. Predicting Outcomes of Antidepressant Treatment in Community Practice Settings. Psychiatr Serv 2024; 75:419-426. [PMID: 38050444 DOI: 10.1176/appi.ps.20230380] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 12/06/2023]
Abstract
OBJECTIVE The authors examined whether machine-learning models could be used to analyze data from electronic health records (EHRs) to predict patients' responses to antidepressant medications. METHODS EHR data from a Washington State health system identified patients ages ≥13 years who started an antidepressant medication in 2016 in a community practice setting and had a baseline Patient Health Questionnaire-9 (PHQ-9) score of ≥10 and at least one PHQ-9 score recorded 14-180 days later. Potential predictors of a response to antidepressants were extracted from the EHR and included demographic characteristics, psychiatric and substance use diagnoses, past psychiatric medication use, mental health service use, and past PHQ-9 scores. Random-forest and penalized regression analyses were used to build models predicting follow-up PHQ-9 score and a favorable treatment response (≥50% improvement in score). RESULTS Among 2,469 patients starting antidepressant medication treatment, the mean±SD baseline PHQ-9 score was 17.3±4.5, and the mean lowest follow-up score was 9.2±5.9. Outcome data were available for 72% of the patients. About 48% of the patients had a favorable treatment response. The best-fitting random-forest models yielded a correlation between predicted and observed follow-up scores of 0.38 (95% CI=0.32-0.45) and an area under the receiver operating characteristic curve for a favorable response of 0.57 (95% CI=0.52-0.61). Results were similar for penalized regression models and for models predicting last PHQ-9 score during follow-up. CONCLUSIONS Prediction models using EHR data were not accurate enough to inform recommendations for or against starting antidepressant medication. Personalization of depression treatment should instead rely on systematic assessment of early outcomes.
Collapse
Affiliation(s)
- Gregory E Simon
- Kaiser Permanente Washington Health Research Institute, Seattle (Simon, Cruz, Shortreed, Coley); Kaiser Permanente Colorado Institute for Health Research, Aurora (Boggs, Beck)
| | - Maricela Cruz
- Kaiser Permanente Washington Health Research Institute, Seattle (Simon, Cruz, Shortreed, Coley); Kaiser Permanente Colorado Institute for Health Research, Aurora (Boggs, Beck)
| | - Jennifer M Boggs
- Kaiser Permanente Washington Health Research Institute, Seattle (Simon, Cruz, Shortreed, Coley); Kaiser Permanente Colorado Institute for Health Research, Aurora (Boggs, Beck)
| | - Arne Beck
- Kaiser Permanente Washington Health Research Institute, Seattle (Simon, Cruz, Shortreed, Coley); Kaiser Permanente Colorado Institute for Health Research, Aurora (Boggs, Beck)
| | - Susan M Shortreed
- Kaiser Permanente Washington Health Research Institute, Seattle (Simon, Cruz, Shortreed, Coley); Kaiser Permanente Colorado Institute for Health Research, Aurora (Boggs, Beck)
| | - R Yates Coley
- Kaiser Permanente Washington Health Research Institute, Seattle (Simon, Cruz, Shortreed, Coley); Kaiser Permanente Colorado Institute for Health Research, Aurora (Boggs, Beck)
| |
Collapse
|
3
|
Nunez JJ, Leung B, Ho C, Ng RT, Bates AT. Predicting which patients with cancer will see a psychiatrist or counsellor from their initial oncology consultation document using natural language processing. COMMUNICATIONS MEDICINE 2024; 4:69. [PMID: 38589545 PMCID: PMC11001970 DOI: 10.1038/s43856-024-00495-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2023] [Accepted: 03/28/2024] [Indexed: 04/10/2024] Open
Abstract
BACKGROUND Patients with cancer often have unmet psychosocial needs. Early detection of who requires referral to a counsellor or psychiatrist may improve their care. This work used natural language processing to predict which patients will see a counsellor or psychiatrist from a patient's initial oncology consultation document. We believe this is the first use of artificial intelligence to predict psychiatric outcomes from non-psychiatric medical documents. METHODS This retrospective prognostic study used data from 47,625 patients at BC Cancer. We analyzed initial oncology consultation documents using traditional and neural language models to predict whether patients would see a counsellor or psychiatrist in the 12 months following their initial oncology consultation. RESULTS Here, we show our best models achieved a balanced accuracy (receiver-operating-characteristic area-under-curve) of 73.1% (0.824) for predicting seeing a psychiatrist, and 71.0% (0.784) for seeing a counsellor. Different words and phrases are important for predicting each outcome. CONCLUSION These results suggest natural language processing can be used to predict psychosocial needs of patients with cancer from their initial oncology consultation document. Future research could extend this work to predict the psychosocial needs of medical patients in other settings.
Collapse
Affiliation(s)
- John-Jose Nunez
- BC Cancer, Vancouver, BC, Canada.
- Department of Computer Science, University of British Columbia, Vancouver, BC, Canada.
- Department of Psychiatry, University of British Columbia, Vancouver, BC, Canada.
| | | | | | - Raymond T Ng
- Department of Computer Science, University of British Columbia, Vancouver, BC, Canada
| | - Alan T Bates
- BC Cancer, Vancouver, BC, Canada
- Department of Psychiatry, University of British Columbia, Vancouver, BC, Canada
| |
Collapse
|
4
|
Wang J, Wu DD, DeLorenzo C, Yang J. Examining factors related to low performance of predicting remission in participants with major depressive disorder using neuroimaging data and other clinical features. PLoS One 2024; 19:e0299625. [PMID: 38547128 PMCID: PMC10977765 DOI: 10.1371/journal.pone.0299625] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2023] [Accepted: 02/13/2024] [Indexed: 04/02/2024] Open
Abstract
Major depressive disorder (MDD), a prevalent mental health issue, affects more than 8% of the US population, and almost 17% in the young group of 18-25 years old. Since Covid-19, its prevalence has become even more significant. However, the remission (being free of depression) rates of first-line antidepressant treatments on MDD are only about 30%. To improve treatment outcomes, researchers have built various predictive models for treatment responses and yet none of them have been adopted in clinical use. One reason is that most predictive models are based on data from subjective questionnaires, which are less reliable. Neuroimaging data are promising objective prognostic factors, but they are expensive to obtain and hence predictive models using neuroimaging data are limited and such studies were usually in small scale (N<100). In this paper, we proposed an advanced machine learning (ML) pipeline for small training dataset with large number of features. We implemented multiple imputation for missing data and repeated K-fold cross validation (CV) to robustly estimate predictive performances. Different feature selection methods and stacking methods using 6 general ML models including random forest, gradient boosting decision tree, XGBoost, penalized logistic regression, support vector machine (SVM), and neural network were examined to evaluate the model performances. All predictive models were compared using model performance metrics such as accuracy, balanced accuracy, area under ROC curve (AUC), sensitivity and specificity. Our proposed ML pipeline was applied to a training dataset and obtained an accuracy and AUC above 0.80. But such high performance failed while applying our ML pipeline using an external validation dataset from the EMBARC study which is a multi-center study. We further examined the possible reasons especially the site heterogeneity issue.
Collapse
Affiliation(s)
- Junying Wang
- Department of Applied Mathematics and Statistics, Stony Brook University, New York, New York, United states of America
| | - David D. Wu
- School of Engineering, University of Michigan, Ann Arbor, Michigan, United States of America
| | - Christine DeLorenzo
- Department of Psychiatry and Behavioral Health, Stony Brook University, Stony Brook, New York, United States of America
- Department of Biomedical Engineering, Stony Brook University, Stony Brook, New York, United States of America
| | - Jie Yang
- Department of Family, Population & Preventive Medicine, Stony Brook University, Stony Brook, New York, United States of America
| |
Collapse
|
5
|
Cai Y, Cai YQ, Tang LY, Wang YH, Gong M, Jing TC, Li HJ, Li-Ling J, Hu W, Yin Z, Gong DX, Zhang GW. Artificial intelligence in the risk prediction models of cardiovascular disease and development of an independent validation screening tool: a systematic review. BMC Med 2024; 22:56. [PMID: 38317226 PMCID: PMC10845808 DOI: 10.1186/s12916-024-03273-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/16/2023] [Accepted: 01/23/2024] [Indexed: 02/07/2024] Open
Abstract
BACKGROUND A comprehensive overview of artificial intelligence (AI) for cardiovascular disease (CVD) prediction and a screening tool of AI models (AI-Ms) for independent external validation are lacking. This systematic review aims to identify, describe, and appraise AI-Ms of CVD prediction in the general and special populations and develop a new independent validation score (IVS) for AI-Ms replicability evaluation. METHODS PubMed, Web of Science, Embase, and IEEE library were searched up to July 2021. Data extraction and analysis were performed for the populations, distribution, predictors, algorithms, etc. The risk of bias was evaluated with the prediction risk of bias assessment tool (PROBAST). Subsequently, we designed IVS for model replicability evaluation with five steps in five items, including transparency of algorithms, performance of models, feasibility of reproduction, risk of reproduction, and clinical implication, respectively. The review is registered in PROSPERO (No. CRD42021271789). RESULTS In 20,887 screened references, 79 articles (82.5% in 2017-2021) were included, which contained 114 datasets (67 in Europe and North America, but 0 in Africa). We identified 486 AI-Ms, of which the majority were in development (n = 380), but none of them had undergone independent external validation. A total of 66 idiographic algorithms were found; however, 36.4% were used only once and only 39.4% over three times. A large number of different predictors (range 5-52,000, median 21) and large-span sample size (range 80-3,660,000, median 4466) were observed. All models were at high risk of bias according to PROBAST, primarily due to the incorrect use of statistical methods. IVS analysis confirmed only 10 models as "recommended"; however, 281 and 187 were "not recommended" and "warning," respectively. CONCLUSION AI has led the digital revolution in the field of CVD prediction, but is still in the early stage of development as the defects of research design, report, and evaluation systems. The IVS we developed may contribute to independent external validation and the development of this field.
Collapse
Affiliation(s)
- Yue Cai
- China Medical University, Shenyang, 110122, China
| | - Yu-Qing Cai
- China Medical University, Shenyang, 110122, China
| | - Li-Ying Tang
- China Medical University, Shenyang, 110122, China
| | - Yi-Han Wang
- China Medical University, Shenyang, 110122, China
| | - Mengchun Gong
- Digital Health China Co. Ltd, Beijing, 100089, China
| | - Tian-Ci Jing
- Smart Hospital Management Department, the First Hospital of China Medical University, Shenyang, 110001, China
| | - Hui-Jun Li
- Shenyang Medical & Film Science and Technology Co. Ltd., Shenyang, 110001, China
- Enduring Medicine Smart Innovation Research Institute, Shenyang, 110001, China
| | - Jesse Li-Ling
- Institute of Genetic Medicine, School of Life Science, State Key Laboratory of Biotherapy, Sichuan University, Chengdu, 610065, China
| | - Wei Hu
- Bayi Orthopedic Hospital, Chengdu, 610017, China
| | - Zhihua Yin
- Department of Epidemiology, School of Public Health, China Medical University, Shenyang, 110122, China.
| | - Da-Xin Gong
- Smart Hospital Management Department, the First Hospital of China Medical University, Shenyang, 110001, China.
- The Internet Hospital Branch of the Chinese Research Hospital Association, Beijing, 100006, China.
| | - Guang-Wei Zhang
- Smart Hospital Management Department, the First Hospital of China Medical University, Shenyang, 110001, China.
- The Internet Hospital Branch of the Chinese Research Hospital Association, Beijing, 100006, China.
| |
Collapse
|
6
|
Sheu YH, Magdamo C, Miller M, Das S, Blacker D, Smoller JW. AI-assisted prediction of differential response to antidepressant classes using electronic health records. NPJ Digit Med 2023; 6:73. [PMID: 37100858 PMCID: PMC10133261 DOI: 10.1038/s41746-023-00817-8] [Citation(s) in RCA: 7] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2022] [Accepted: 04/04/2023] [Indexed: 04/28/2023] Open
Abstract
Antidepressant selection is largely a trial-and-error process. We used electronic health record (EHR) data and artificial intelligence (AI) to predict response to four antidepressants classes (SSRI, SNRI, bupropion, and mirtazapine) 4 to 12 weeks after antidepressant initiation. The final data set comprised 17,556 patients. Predictors were derived from both structured and unstructured EHR data and models accounted for features predictive of treatment selection to minimize confounding by indication. Outcome labels were derived through expert chart review and AI-automated imputation. Regularized generalized linear model (GLM), random forest, gradient boosting machine (GBM), and deep neural network (DNN) models were trained and their performance compared. Predictor importance scores were derived using SHapley Additive exPlanations (SHAP). All models demonstrated similarly good prediction performance (AUROCs ≥ 0.70, AUPRCs ≥ 0.68). The models can estimate differential treatment response probabilities both between patients and between antidepressant classes for the same patient. In addition, patient-specific factors driving response probabilities for each antidepressant class can be generated. We show that antidepressant response can be accurately predicted from real-world EHR data with AI modeling, and our approach could inform further development of clinical decision support systems for more effective treatment selection.
Collapse
Affiliation(s)
- Yi-Han Sheu
- Center for Precision Psychiatry, Massachusetts General Hospital, Boston, MA, USA.
- Psychiatric and Neurodevelopmental Genetics Unit, Center for Genomic Medicine, Massachusetts General Hospital, Boston, MA, USA.
- Department of Psychiatry, Massachusetts General Hospital / Harvard Medical School, Boston, MA, USA.
- Stanley Center for Psychiatric Research, Broad Institute of MIT and Harvard, Cambridge, MA, USA.
| | - Colin Magdamo
- Department of Neurology, Massachusetts General Hospital / Harvard Medical School, Boston, MA, USA
| | - Matthew Miller
- Harvard Injury Control Research Center, Harvard T.H. Chan School of Public Health, Boston, MA, USA
- Bouvé College of Health Sciences, Northeastern University, Boston, MA, USA
- Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, MA, USA
| | - Sudeshna Das
- Department of Neurology, Massachusetts General Hospital / Harvard Medical School, Boston, MA, USA
| | - Deborah Blacker
- Department of Psychiatry, Massachusetts General Hospital / Harvard Medical School, Boston, MA, USA
- Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, MA, USA
| | - Jordan W Smoller
- Center for Precision Psychiatry, Massachusetts General Hospital, Boston, MA, USA
- Psychiatric and Neurodevelopmental Genetics Unit, Center for Genomic Medicine, Massachusetts General Hospital, Boston, MA, USA
- Department of Psychiatry, Massachusetts General Hospital / Harvard Medical School, Boston, MA, USA
- Stanley Center for Psychiatric Research, Broad Institute of MIT and Harvard, Cambridge, MA, USA
- Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, MA, USA
| |
Collapse
|
7
|
Automated diagnosis of depression from EEG signals using traditional and deep learning approaches: A comparative analysis. Biocybern Biomed Eng 2022. [DOI: 10.1016/j.bbe.2021.12.005] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/20/2022]
|