1
|
Kim S, Kang U, Gu J, Kim J, Park J, Hwang GW, Park S, Jang HJ, Seong TY, Lee S. Artificial Multimodal Neuron with Associative Learning Capabilities: Acquisition, Extinction, and Spontaneous Recovery. ACS APPLIED MATERIALS & INTERFACES 2024; 16:36519-36526. [PMID: 38950119 DOI: 10.1021/acsami.4c02343] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/03/2024]
Abstract
Associative multimodal artificial intelligence (AMAI) has gained significant attention across various fields, yet its implementation poses challenges due to the burden on computing and memory resources. To address these challenges, researchers have paid increasing attention to neuromorphic devices based on novel materials and structures, which can implement classical conditioning behaviors with simplified circuitry. Herein, we introduce an artificial multimodal neuron device that shows not only the acquisition behavior but also the extinction and the spontaneous recovery behaviors for the first time. Being composed of an ovonic threshold switch (OTS)-based neuron device, a conductive bridge memristor (CBM)-based synapse device, and a few passive electrical elements, such observed behaviors of this neuron device are explained in terms of the electroforming and the diffusion of metallic ions in the CBM. We believe that the proposed associative learning neuron device will shed light on the way of developing large-scale AMAI systems by providing inspiration to devise an associative learning network with improved energy efficiency.
Collapse
Affiliation(s)
- Sangheon Kim
- Center for Neuromorphic Engineering, Korea Institute of Science and Technology, Seoul 02792, Rep. of Korea
- Department of Materials Science and Engineering, Korea University, Seoul 02841, Rep. of Korea
| | - Unhyeon Kang
- Center for Neuromorphic Engineering, Korea Institute of Science and Technology, Seoul 02792, Rep. of Korea
- Materials Science & Engineering, Seoul National University, Seoul 08826, Rep. of Korea
| | - Jiyoung Gu
- Center for Neuromorphic Engineering, Korea Institute of Science and Technology, Seoul 02792, Rep. of Korea
- Department of Materials Science & Engineering, Seoul National University of Science and Technology, Seoul 01811, Rep. of Korea
| | - Jaewook Kim
- Center for Neuromorphic Engineering, Korea Institute of Science and Technology, Seoul 02792, Rep. of Korea
| | - Jongkil Park
- Center for Neuromorphic Engineering, Korea Institute of Science and Technology, Seoul 02792, Rep. of Korea
| | - Gyu Weon Hwang
- Center for Neuromorphic Engineering, Korea Institute of Science and Technology, Seoul 02792, Rep. of Korea
| | - Seongsik Park
- Center for Neuromorphic Engineering, Korea Institute of Science and Technology, Seoul 02792, Rep. of Korea
| | - Hyun Jae Jang
- Center for Neuromorphic Engineering, Korea Institute of Science and Technology, Seoul 02792, Rep. of Korea
| | - Tae-Yeon Seong
- Department of Materials Science and Engineering, Korea University, Seoul 02841, Rep. of Korea
| | - Suyoun Lee
- Center for Neuromorphic Engineering, Korea Institute of Science and Technology, Seoul 02792, Rep. of Korea
- Division of Nano & Information Technology, Korea University of Science and Technology, Daejon 34316, Rep. of Korea
| |
Collapse
|
2
|
Feng R, Li S, Zhang Y. AI-powered microscopy image analysis for parasitology: integrating human expertise. Trends Parasitol 2024; 40:633-646. [PMID: 38824067 DOI: 10.1016/j.pt.2024.05.005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2024] [Revised: 05/06/2024] [Accepted: 05/07/2024] [Indexed: 06/03/2024]
Abstract
Microscopy image analysis plays a pivotal role in parasitology research. Deep learning (DL), a subset of artificial intelligence (AI), has garnered significant attention. However, traditional DL-based methods for general purposes are data-driven, often lacking explainability due to their black-box nature and sparse instructional resources. To address these challenges, this article presents a comprehensive review of recent advancements in knowledge-integrated DL models tailored for microscopy image analysis in parasitology. The massive amounts of human expert knowledge from parasitologists can enhance the accuracy and explainability of AI-driven decisions. It is expected that the adoption of knowledge-integrated DL models will open up a wide range of applications in the field of parasitology.
Collapse
Affiliation(s)
- Ruijun Feng
- College of Science, Harbin Institute of Technology (Shenzhen), Shenzhen, Guangdong 518055, China; School of Computer Science and Engineering, University of New South Wales, Sydney, Australia
| | - Sen Li
- College of Science, Harbin Institute of Technology (Shenzhen), Shenzhen, Guangdong 518055, China
| | - Yang Zhang
- College of Science, Harbin Institute of Technology (Shenzhen), Shenzhen, Guangdong 518055, China.
| |
Collapse
|
3
|
Parsa S, Somani S, Dudum R, Jain SS, Rodriguez F. Artificial Intelligence in Cardiovascular Disease Prevention: Is it Ready for Prime Time? Curr Atheroscler Rep 2024; 26:263-272. [PMID: 38780665 DOI: 10.1007/s11883-024-01210-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 05/08/2024] [Indexed: 05/25/2024]
Abstract
PURPOSE OF REVIEW This review evaluates how Artificial Intelligence (AI) enhances atherosclerotic cardiovascular disease (ASCVD) risk assessment, allows for opportunistic screening, and improves adherence to guidelines through the analysis of unstructured clinical data and patient-generated data. Additionally, it discusses strategies for integrating AI into clinical practice in preventive cardiology. RECENT FINDINGS AI models have shown superior performance in personalized ASCVD risk evaluations compared to traditional risk scores. These models now support automated detection of ASCVD risk markers, including coronary artery calcium (CAC), across various imaging modalities such as dedicated ECG-gated CT scans, chest X-rays, mammograms, coronary angiography, and non-gated chest CT scans. Moreover, large language model (LLM) pipelines are effective in identifying and addressing gaps and disparities in ASCVD preventive care, and can also enhance patient education. AI applications are proving invaluable in preventing and managing ASCVD and are primed for clinical use, provided they are implemented within well-regulated, iterative clinical pathways.
Collapse
Affiliation(s)
- Shyon Parsa
- Department of Medicine, Stanford University, Stanford, California, USA
| | - Sulaiman Somani
- Department of Medicine, Stanford University, Stanford, California, USA
| | - Ramzi Dudum
- Division of Cardiovascular Medicine and Cardiovascular Institute, Stanford University, Stanford, CA, USA
| | - Sneha S Jain
- Division of Cardiovascular Medicine and Cardiovascular Institute, Stanford University, Stanford, CA, USA
| | - Fatima Rodriguez
- Division of Cardiovascular Medicine and Cardiovascular Institute, Stanford University, Stanford, CA, USA.
- Center for Digital Health, Stanford University, Stanford, California, USA.
| |
Collapse
|
4
|
Lee SY, Lee DY, Ahn J. Evaluation of machine learning approach for surgical results of Ahmed valve implantation in patients with glaucoma. BMC Ophthalmol 2024; 24:248. [PMID: 38862946 PMCID: PMC11167936 DOI: 10.1186/s12886-024-03510-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2024] [Accepted: 06/03/2024] [Indexed: 06/13/2024] Open
Abstract
BACKGROUND Ahmed valve implantation demonstrated an increasing proportion in glaucoma surgery, but predicting the successful maintenance of target intraocular pressure remains a challenging task. This study aimed to evaluate the performance of machine learning (ML) in predicting surgical outcomes after Ahmed valve implantation and to assess potential risk factors associated with surgical failure to contribute to improving the success rate. METHODS This study used preoperative data of patients who underwent Ahmed valve implantation from 2017 to 2021 at Ajou University Hospital. These datasets included demographic and ophthalmic parameters (dataset A), systemic medical records excluding psychiatric records (dataset B), and psychiatric medications (dataset C). Logistic regression, extreme gradient boosting (XGBoost), and support vector machines were first evaluated using only dataset A. The algorithm with the best performance was selected based on the area under the receiver operating characteristics curve (AUROC). Finally, three additional prediction models were developed using the best performance algorithm, incorporating combinations of multiple datasets to predict surgical outcomes at 1 year. RESULTS Among 153 eyes of 133 patients, 131 (85.6%) and 22 (14.4%) eyes were categorized as the success and failure groups, respectively. The XGBoost was shown as the best-performance model with an AUROC value of 0.684, using only dataset A. The final three further prediction models were developed based on the combination of multiple datasets using the XGBoost model. All datasets combinations demonstrated the best performances in terms of AUROC (dataset A + B: 0.782; A + C: 0.773; A + B + C: 0.801). Furthermore, advancing age was a risk factor associated with a higher surgical failure incidence. CONCLUSIONS ML provides some predictive value in predicting the outcomes of Ahmed valve implantation at 1 year. ML evaluation revealed advancing age as a common risk factor for surgical failure.
Collapse
Affiliation(s)
- Seung Yeop Lee
- Department of Ophthalmology, Ajou University Medical Center, Ajou University School of Medicine, 164, World Cup-ro, Yeongtong-gu, Suwon, 16499, Republic of Korea
| | - Dong Yun Lee
- Department of Biomedical Informatics, Ajou University School of Medicine, 154, Word Cup-ro, Yeongtong-gu, Suwon, 16499, Republic of Korea
| | - Jaehong Ahn
- Department of Ophthalmology, Ajou University Medical Center, Ajou University School of Medicine, 164, World Cup-ro, Yeongtong-gu, Suwon, 16499, Republic of Korea.
| |
Collapse
|
5
|
Baheti B, Innani S, Nasrallah M, Bakas S. Prognostic stratification of glioblastoma patients by unsupervised clustering of morphology patterns on whole slide images furthering our disease understanding. Front Neurosci 2024; 18:1304191. [PMID: 38831756 PMCID: PMC11146603 DOI: 10.3389/fnins.2024.1304191] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2023] [Accepted: 04/25/2024] [Indexed: 06/05/2024] Open
Abstract
Introduction Glioblastoma (GBM) is a highly aggressive malignant tumor of the central nervous system that displays varying molecular and morphological profiles, leading to challenging prognostic assessments. Stratifying GBM patients according to overall survival (OS) from H&E-stained whole slide images (WSI) using advanced computational methods is challenging, but with direct clinical implications. Methods This work is focusing on GBM (IDH-wildtype, CNS WHO Gr.4) cases, identified from the TCGA-GBM and TCGA-LGG collections after considering the 2021 WHO classification criteria. The proposed approach starts with patch extraction in each WSI, followed by comprehensive patch-level curation to discard artifactual content, i.e., glass reflections, pen markings, dust on the slide, and tissue tearing. Each patch is then computationally described as a feature vector defined by a pre-trained VGG16 convolutional neural network. Principal component analysis provides a feature representation of reduced dimensionality, further facilitating identification of distinct groups of morphology patterns, via unsupervised k-means clustering. Results The optimal number of clusters, according to cluster reproducibility and separability, is automatically determined based on the rand index and silhouette coefficient, respectively. Our proposed approach achieved prognostic stratification accuracy of 83.33% on a multi-institutional independent unseen hold-out test set with sensitivity and specificity of 83.33%. Discussion We hypothesize that the quantification of these clusters of morphology patterns, reflect the tumor's spatial heterogeneity and yield prognostic relevant information to distinguish between short and long survivors using a decision tree classifier. The interpretability analysis of the obtained results can contribute to furthering and quantifying our understanding of GBM and potentially improving our diagnostic and prognostic predictions.
Collapse
Affiliation(s)
- Bhakti Baheti
- Division of Computational Pathology, Department of Pathology and Laboratory Medicine, Indiana University School of Medicine, Indianapolis, IN, United States
- Center for Artificial Intelligence and Data Science for Integrated Diagnostics (AI2D) and Center for Biomedical Image Computing and Analytics (CBICA), University of Pennsylvania, Philadelphia, PA, United States
- Department of Pathology and Laboratory Medicine, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, United States
| | - Shubham Innani
- Division of Computational Pathology, Department of Pathology and Laboratory Medicine, Indiana University School of Medicine, Indianapolis, IN, United States
- Center for Artificial Intelligence and Data Science for Integrated Diagnostics (AI2D) and Center for Biomedical Image Computing and Analytics (CBICA), University of Pennsylvania, Philadelphia, PA, United States
- Department of Pathology and Laboratory Medicine, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, United States
| | - MacLean Nasrallah
- Center for Artificial Intelligence and Data Science for Integrated Diagnostics (AI2D) and Center for Biomedical Image Computing and Analytics (CBICA), University of Pennsylvania, Philadelphia, PA, United States
- Department of Pathology and Laboratory Medicine, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, United States
| | - Spyridon Bakas
- Division of Computational Pathology, Department of Pathology and Laboratory Medicine, Indiana University School of Medicine, Indianapolis, IN, United States
- Center for Artificial Intelligence and Data Science for Integrated Diagnostics (AI2D) and Center for Biomedical Image Computing and Analytics (CBICA), University of Pennsylvania, Philadelphia, PA, United States
- Department of Pathology and Laboratory Medicine, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, United States
- Department of Radiology, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, United States
- Department of Computer Science, Luddy School of Informatics, Computing, and Engineering, Indiana University, Indianapolis, IN, United States
| |
Collapse
|
6
|
Lopes J, Faria M, Santos MF. Exploring trends and autonomy levels of adaptive business intelligence in healthcare: A systematic review. PLoS One 2024; 19:e0302697. [PMID: 38728308 PMCID: PMC11086907 DOI: 10.1371/journal.pone.0302697] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/24/2023] [Accepted: 04/09/2024] [Indexed: 05/12/2024] Open
Abstract
OBJECTIVE In order to comprehensively understand the characteristics of Adaptive Business Intelligence (ABI) in Healthcare, this study is structured to provide insights into the common features and evolving patterns within this domain. Applying the Sheridan's Classification as a framework, we aim to assess the degree of autonomy exhibited by various ABI components. Together, these objectives will contribute to a deeper understanding of ABI implementation and its implications within the Healthcare context. METHODS A comprehensive search of academic databases was conducted to identify relevant studies, selecting AIS e-library (AISel), Decision Support Systems Journal (DSSJ), Nature, The Lancet Digital Health (TLDH), PubMed, Expert Systems with Application (ESWA) and npj Digital Medicine as information sources. Studies from 2006 to 2022 were included based on predefined eligibility criteria. PRISMA statements were used to report this study. RESULTS The outcomes showed that ABI systems present distinct levels of development, autonomy and practical deployment. The high levels of autonomy were essentially associated with predictive components. However, the possibility of completely autonomous decisions by these systems is totally excluded. Lower levels of autonomy are also observed, particularly in connection with prescriptive components, granting users responsibility in the generation of decisions. CONCLUSION The study presented emphasizes the vital connection between desired outcomes and the inherent autonomy of these solutions, highlighting the critical need for additional research on the consequences of ABI systems and their constituent elements. Organizations should deploy these systems in a way consistent with their objectives and values, while also being mindful of potential adverse effects. Providing valuable insights for researchers, practitioners, and policymakers aiming to comprehend the diverse levels of ABI systems implementation, it contributes to well-informed decision-making in this dynamic field.
Collapse
Affiliation(s)
- João Lopes
- ALGORITMI Research Center, University of Minho, Braga, Portugal
| | - Mariana Faria
- ALGORITMI Research Center, University of Minho, Braga, Portugal
| | | |
Collapse
|
7
|
Tamura Y, Nomura A, Kagiyama N, Mizuno A, Node K. Digitalomics, digital intervention, and designing future: The next frontier in cardiology. J Cardiol 2024; 83:318-322. [PMID: 38135148 DOI: 10.1016/j.jjcc.2023.12.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/02/2023] [Revised: 12/10/2023] [Accepted: 12/15/2023] [Indexed: 12/24/2023]
Abstract
The discipline of cardiology stands at a transformative juncture, primarily influenced by the surge in digital health technologies. These innovations hold the promise to redefine the realms of cardiovascular research and patient care, ushering in an era of individualized and data-driven treatments. This review delves into the heart of this evolution, introducing a comprehensive design for the future of cardiology. Emphasizing the emerging domains of "digitalomics" and "digital intervention", it explores how the integration of patient data, artificial intelligence-enabled diagnostics, and telehealth can lead to more streamlined and personalized cardiovascular health. The "digital-twin" model, a highlight of this approach, encapsulates individual patient characteristics, allowing for targeted treatments. The role of interdisciplinary collaboration in cardiovascular medicine is also underlined, emphasizing the importance of merging traditional cardiology with technological advancements. The convergence of traditional cardiology methods and digital health technologies, facilitated by a transdisciplinary approach, is set to chart a new course in cardiovascular health, emphasizing individualized care and improved clinical outcomes.
Collapse
Affiliation(s)
- Yuichi Tamura
- Pulmonary Hypertension Center, International University of Health and Welfare Mita Hospital, Tokyo, Japan; Department of Cardiology International University of Health and Welfare School of Medicine Narita, Japan; Cardiointelligence Inc., Tokyo, Japan.
| | - Akihiro Nomura
- College of Transdisciplinary Sciences for Innovation, Kanazawa University, Kanazawa, Japan; Department of Cardiovascular Medicine, Kanazawa University Graduate School of Medical Sciences, Kanazawa, Japan; Frontier Institute of Tourism Sciences, Kanazawa University, Kanazawa, Japan; Department of Biomedical Informatics, CureApp Institute, Karuizawa, Japan
| | - Nobuyuki Kagiyama
- Department of Digital Health and Telemedicine R&D, Juntendo University, Tokyo, Japan; Department of Cardiovascular Biology and Medicine, Juntendo University Graduate School of Medicine, Tokyo, Japan
| | - Atsushi Mizuno
- Department of Cardiovascular Medicine, St. Luke's International Hospital, Tokyo, Japan; Leonard Davis Institute for Health Economics, University of Pennsylvania, PA, USA
| | - Koichi Node
- Department of Cardiovascular Medicine, Saga University, Saga, Japan
| |
Collapse
|
8
|
Sujit SJ, Aminu M, Karpinets TV, Chen P, Saad MB, Salehjahromi M, Boom JD, Qayati M, George JM, Allen H, Antonoff MB, Hong L, Hu X, Heeke S, Tran HT, Le X, Elamin YY, Altan M, Vokes NI, Sheshadri A, Lin J, Zhang J, Lu Y, Behrens C, Godoy MCB, Wu CC, Chang JY, Chung C, Jaffray DA, Wistuba II, Lee JJ, Vaporciyan AA, Gibbons DL, Heymach J, Zhang J, Cascone T, Wu J. Enhancing NSCLC recurrence prediction with PET/CT habitat imaging, ctDNA, and integrative radiogenomics-blood insights. Nat Commun 2024; 15:3152. [PMID: 38605064 PMCID: PMC11009351 DOI: 10.1038/s41467-024-47512-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2023] [Accepted: 03/21/2024] [Indexed: 04/13/2024] Open
Abstract
While we recognize the prognostic importance of clinicopathological measures and circulating tumor DNA (ctDNA), the independent contribution of quantitative image markers to prognosis in non-small cell lung cancer (NSCLC) remains underexplored. In our multi-institutional study of 394 NSCLC patients, we utilize pre-treatment computed tomography (CT) and 18F-fluorodeoxyglucose positron emission tomography (FDG-PET) to establish a habitat imaging framework for assessing regional heterogeneity within individual tumors. This framework identifies three PET/CT subtypes, which maintain prognostic value after adjusting for clinicopathologic risk factors including tumor volume. Additionally, these subtypes complement ctDNA in predicting disease recurrence. Radiogenomics analysis unveil the molecular underpinnings of these imaging subtypes, highlighting downregulation in interferon alpha and gamma pathways in the high-risk subtype. In summary, our study demonstrates that these habitat imaging subtypes effectively stratify NSCLC patients based on their risk levels for disease recurrence after initial curative surgery or radiotherapy, providing valuable insights for personalized treatment approaches.
Collapse
Affiliation(s)
- Sheeba J Sujit
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Muhammad Aminu
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Tatiana V Karpinets
- Department of Genomic Medicine, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Pingjun Chen
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Maliazurina B Saad
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Morteza Salehjahromi
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - John D Boom
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
- Department of Biomedical Engineering, Duke University, Durham, NC, USA
| | - Mohamed Qayati
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - James M George
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
- Michigan Center for Translational Pathology, University of Michigan, Ann Arbor, MI, USA
| | - Haley Allen
- Natural Sciences, Rice University, Houston, TX, USA
| | - Mara B Antonoff
- Department of Thoracic and Cardiovascular Surgery, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Lingzhi Hong
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Xin Hu
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Simon Heeke
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Hai T Tran
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Xiuning Le
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Yasir Y Elamin
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Mehmet Altan
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Natalie I Vokes
- Department of Genomic Medicine, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Ajay Sheshadri
- Department of Pulmonary Medicine, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Julie Lin
- Department of Pulmonary Medicine, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Jianhua Zhang
- Department of Genomic Medicine, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Yang Lu
- Department of Nuclear Medicine, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Carmen Behrens
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Myrna C B Godoy
- Department of Thoracic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Carol C Wu
- Department of Thoracic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Joe Y Chang
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Caroline Chung
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
- Institute of Data Science in Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - David A Jaffray
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
- Institute of Data Science in Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Ignacio I Wistuba
- Department of Translational Molecular Pathology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - J Jack Lee
- Department of Biostatistics, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Ara A Vaporciyan
- Department of Thoracic and Cardiovascular Surgery, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Don L Gibbons
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - John Heymach
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Jianjun Zhang
- Department of Genomic Medicine, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
- Lung Cancer Genomics Program, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
- Lung Cancer Interception Program, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Tina Cascone
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
| | - Jia Wu
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, USA.
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA.
- Institute of Data Science in Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA.
| |
Collapse
|
9
|
Muse ED, Topol EJ. Transforming the cardiometabolic disease landscape: Multimodal AI-powered approaches in prevention and management. Cell Metab 2024; 36:670-683. [PMID: 38428435 PMCID: PMC10990799 DOI: 10.1016/j.cmet.2024.02.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/26/2023] [Revised: 01/25/2024] [Accepted: 02/06/2024] [Indexed: 03/03/2024]
Abstract
The rise of artificial intelligence (AI) has revolutionized various scientific fields, particularly in medicine, where it has enabled the modeling of complex relationships from massive datasets. Initially, AI algorithms focused on improved interpretation of diagnostic studies such as chest X-rays and electrocardiograms in addition to predicting patient outcomes and future disease onset. However, AI has evolved with the introduction of transformer models, allowing analysis of the diverse, multimodal data sources existing in medicine today. Multimodal AI holds great promise in more accurate disease risk assessment and stratification as well as optimizing the key driving factors in cardiometabolic disease: blood pressure, sleep, stress, glucose control, weight, nutrition, and physical activity. In this article we outline the current state of medical AI in cardiometabolic disease, highlighting the potential of multimodal AI to augment personalized prevention and treatment strategies in cardiometabolic disease.
Collapse
Affiliation(s)
- Evan D Muse
- Scripps Research Translational Institute, Scripps Research, La Jolla, CA 92037, USA; Division of Cardiovascular Diseases, Scripps Clinic, La Jolla, CA 92037, USA
| | - Eric J Topol
- Scripps Research Translational Institute, Scripps Research, La Jolla, CA 92037, USA; Division of Cardiovascular Diseases, Scripps Clinic, La Jolla, CA 92037, USA.
| |
Collapse
|
10
|
Park WY, Jeon K, Schmidt TS, Kondylakis H, Alkasab T, Dewey BE, You SC, Nagy P. Development of Medical Imaging Data Standardization for Imaging-Based Observational Research: OMOP Common Data Model Extension. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:899-908. [PMID: 38315345 PMCID: PMC11031512 DOI: 10.1007/s10278-024-00982-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/04/2023] [Revised: 11/10/2023] [Accepted: 11/14/2023] [Indexed: 02/07/2024]
Abstract
The rapid growth of artificial intelligence (AI) and deep learning techniques require access to large inter-institutional cohorts of data to enable the development of robust models, e.g., targeting the identification of disease biomarkers and quantifying disease progression and treatment efficacy. The Observational Medical Outcomes Partnership Common Data Model (OMOP CDM) has been designed to accommodate a harmonized representation of observational healthcare data. This study proposes the Medical Imaging CDM (MI-CDM) extension, adding two new tables and two vocabularies to the OMOP CDM to address the structural and semantic requirements to support imaging research. The tables provide the capabilities of linking DICOM data sources as well as tracking the provenance of imaging features derived from those images. The implementation of the extension enables phenotype definitions using imaging features and expanding standardized computable imaging biomarkers. This proposal offers a comprehensive and unified approach for conducting imaging research and outcome studies utilizing imaging features.
Collapse
Affiliation(s)
- Woo Yeon Park
- Biomedical Informatics and Data Science, Johns Hopkins University, 855 N Wolfe St, Rangos 616, Baltimore, MD, USA.
| | - Kyulee Jeon
- Department of Biomedical Systems Informatics, Yonsei University College of Medicine, Seoul, Korea
- Institute for Innovation in Digital Healthcare, Yonsei University, Seoul, Korea
| | - Teri Sippel Schmidt
- Biomedical Informatics and Data Science, Johns Hopkins University, 855 N Wolfe St, Rangos 616, Baltimore, MD, USA
| | - Haridimos Kondylakis
- Institute of Computer Science, Foundation of Research & Technology-Hellas (FORTH), Heraklion, Greece
| | - Tarik Alkasab
- Department of Radiology, Massachusetts General Hospital, Boston, MA, USA
| | - Blake E Dewey
- Department of Neurology, Johns Hopkins University, Baltimore, MD, USA
| | - Seng Chan You
- Department of Biomedical Systems Informatics, Yonsei University College of Medicine, Seoul, Korea
- Institute for Innovation in Digital Healthcare, Yonsei University, Seoul, Korea
| | - Paul Nagy
- Biomedical Informatics and Data Science, Johns Hopkins University, 855 N Wolfe St, Rangos 616, Baltimore, MD, USA
| |
Collapse
|
11
|
Zhou Y, Cosentino J, Yun T, Biradar MI, Shreibati J, Lai D, Schwantes-An TH, Luben R, McCaw Z, Engmann J, Providencia R, Schmidt AF, Munroe P, Yang H, Carroll A, Khawaja AP, McLean CY, Behsaz B, Hormozdiari F. Utilizing multimodal AI to improve genetic analyses of cardiovascular traits. MEDRXIV : THE PREPRINT SERVER FOR HEALTH SCIENCES 2024:2024.03.19.24304547. [PMID: 38562791 PMCID: PMC10984061 DOI: 10.1101/2024.03.19.24304547] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/04/2024]
Abstract
Electronic health records, biobanks, and wearable biosensors contain multiple high-dimensional clinical data (HDCD) modalities (e.g., ECG, Photoplethysmography (PPG), and MRI) for each individual. Access to multimodal HDCD provides a unique opportunity for genetic studies of complex traits because different modalities relevant to a single physiological system (e.g., circulatory system) encode complementary and overlapping information. We propose a novel multimodal deep learning method, M-REGLE, for discovering genetic associations from a joint representation of multiple complementary HDCD modalities. We showcase the effectiveness of this model by applying it to several cardiovascular modalities. M-REGLE jointly learns a lower representation (i.e., latent factors) of multimodal HDCD using a convolutional variational autoencoder, performs genome wide association studies (GWAS) on each latent factor, then combines the results to study the genetics of the underlying system. To validate the advantages of M-REGLE and multimodal learning, we apply it to common cardiovascular modalities (PPG and ECG), and compare its results to unimodal learning methods in which representations are learned from each data modality separately, but the downstream genetic analyses are performed on the combined unimodal representations. M-REGLE identifies 19.3% more loci on the 12-lead ECG dataset, 13.0% more loci on the ECG lead I + PPG dataset, and its genetic risk score significantly outperforms the unimodal risk score at predicting cardiac phenotypes, such as atrial fibrillation (Afib), in multiple biobanks.
Collapse
Affiliation(s)
| | | | | | - Mahantesh I Biradar
- NIHR Biomedical Research Centre at Moorfields Eye Hospital & UCL Institute of Ophthalmology, London EC1V 9EL, UK
- MRC Epidemiology Unit, University of Cambridge, Cambridge CB2 0SL, UK
| | | | - Dongbing Lai
- Department of Medical and Molecular Genetics, Indiana University School of Medicine, Indianapolis, IN 46202, USA
| | - Tae-Hwi Schwantes-An
- Department of Medical and Molecular Genetics, Indiana University School of Medicine, Indianapolis, IN 46202, USA
| | - Robert Luben
- NIHR Biomedical Research Centre at Moorfields Eye Hospital & UCL Institute of Ophthalmology, London EC1V 9EL, UK
- MRC Epidemiology Unit, University of Cambridge, Cambridge CB2 0SL, UK
| | - Zachary McCaw
- Department of Biostatistics, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Jorgen Engmann
- Center for Translational Genomics, Population Science and Experimental Medicine, Institute of Cardiovascular Science, University College London, UK
| | - Rui Providencia
- Institute of Health Informatics Research, University College London, London, UK
- Electrophysiology Department, Barts Heart Centre, St. Bartholomew's Hospital, London, UK
| | - Amand Floriaan Schmidt
- Department of Cardiology; Amsterdam University Medical Centres, Amsterdam, The Netherlands
- Institute of Cardiovascular Science; University College London, London, UK
- Division of Heart and Lungs, University Medical Center Utrecht, Utrecht, Netherlands
| | - Patricia Munroe
- William Harvey Research Institute, Barts and the London Faculty of Medicine and Dentistry, Queen Mary University of London, London, UK
| | - Howard Yang
- Google Research, San Francisco CA, 94105 USA
| | | | - Anthony P Khawaja
- NIHR Biomedical Research Centre at Moorfields Eye Hospital & UCL Institute of Ophthalmology, London EC1V 9EL, UK
- MRC Epidemiology Unit, University of Cambridge, Cambridge CB2 0SL, UK
| | | | | | | |
Collapse
|
12
|
Wang Y, Yin C, Zhang P. Multimodal risk prediction with physiological signals, medical images and clinical notes. Heliyon 2024; 10:e26772. [PMID: 38455585 PMCID: PMC10918115 DOI: 10.1016/j.heliyon.2024.e26772] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2023] [Revised: 02/17/2024] [Accepted: 02/20/2024] [Indexed: 03/09/2024] Open
Abstract
The broad adoption of electronic health record (EHR) systems brings us a tremendous amount of clinical data and thus provides opportunities to conduct data-based healthcare research to solve various clinical problems in the medical domain. Machine learning and deep learning methods are widely used in the medical informatics and healthcare domain due to their power to mine insights from raw data. When adapting deep learning models for EHR data, it is essential to consider its heterogeneous nature: EHR contains patient records from various sources including medical tests (e.g. blood test, microbiology test), medical imaging, diagnosis, medications, procedures, clinical notes, etc. Those modalities together provide a holistic view of patient health status and complement each other. Therefore, combining data from multiple modalities that are intrinsically different is challenging but intuitively promising in deep learning for EHR. To assess the expectations of multimodal data, we introduce a comprehensive fusion framework designed to integrate temporal variables, medical images, and clinical notes in EHR for enhanced performance in clinical risk prediction. Early, joint, and late fusion strategies are employed to combine data from various modalities effectively. We test the model with three predictive tasks: in-hospital mortality, long length of stay, and 30-day readmission. Experimental results show that multimodal models outperform uni-modal models in the tasks involved. Additionally, by training models with different input modality combinations, we calculate the Shapley value for each modality to quantify their contribution to multimodal performance. It is shown that temporal variables tend to be more helpful than CXR images and clinical notes in the three explored predictive tasks.
Collapse
Affiliation(s)
- Yuanlong Wang
- Department of Computer Science and Engineering, The Ohio State University, Columbus, OH 43210, USA
| | - Changchang Yin
- Department of Computer Science and Engineering, The Ohio State University, Columbus, OH 43210, USA
- Department of Biomedical Informatics, The Ohio State University, Columbus, OH 43210, USA
| | - Ping Zhang
- Department of Computer Science and Engineering, The Ohio State University, Columbus, OH 43210, USA
- Department of Biomedical Informatics, The Ohio State University, Columbus, OH 43210, USA
| |
Collapse
|
13
|
Lin J, Yang J, Yin M, Tang Y, Chen L, Xu C, Zhu S, Gao J, Liu L, Liu X, Gu C, Huang Z, Wei Y, Zhu J. Development and Validation of Multimodal Models to Predict the 30-Day Mortality of ICU Patients Based on Clinical Parameters and Chest X-Rays. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024:10.1007/s10278-024-01066-1. [PMID: 38448758 DOI: 10.1007/s10278-024-01066-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/06/2023] [Revised: 02/21/2024] [Accepted: 02/22/2024] [Indexed: 03/08/2024]
Abstract
We aimed to develop and validate multimodal ICU patient prognosis models that combine clinical parameters data and chest X-ray (CXR) images. A total of 3798 subjects with clinical parameters and CXR images were extracted from the Medical Information Mart for Intensive Care IV (MIMIC-IV) database and an external hospital (the test set). The primary outcome was 30-day mortality after ICU admission. Automated machine learning (AutoML) and convolutional neural networks (CNNs) were used to construct single-modal models based on clinical parameters and CXR separately. An early fusion approach was used to integrate both modalities (clinical parameters and CXR) into a multimodal model named PrismICU. Compared to the single-modal models, i.e., the clinical parameter model (AUC = 0.80, F1-score = 0.43) and the CXR model (AUC = 0.76, F1-score = 0.45) and the scoring system APACHE II (AUC = 0.83, F1-score = 0.77), PrismICU (AUC = 0.95, F1 score = 0.95) showed improved performance in predicting the 30-day mortality in the validation set. In the test set, PrismICU (AUC = 0.82, F1-score = 0.61) was also better than the clinical parameters model (AUC = 0.72, F1-score = 0.50), CXR model (AUC = 0.71, F1-score = 0.36), and APACHE II (AUC = 0.62, F1-score = 0.50). PrismICU, which integrated clinical parameters data and CXR images, performed better than single-modal models and the existing scoring system. It supports the potential of multimodal models based on structured data and imaging in clinical management.
Collapse
Affiliation(s)
- Jiaxi Lin
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, 188 Shizi Street, Jiangsu, Suzhou 215006, China
- Suzhou Clinical Center of Digestive Diseases, Suzhou, China
| | - Jin Yang
- Department of Critical Care Medicine, The First Affiliated Hospital of Soochow University, 188 Shizi Street, Jiangsu, Suzhou 215006, China
| | - Minyue Yin
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, 188 Shizi Street, Jiangsu, Suzhou 215006, China
- Suzhou Clinical Center of Digestive Diseases, Suzhou, China
| | - Yuxiu Tang
- Department of Critical Care Medicine, The First Affiliated Hospital of Soochow University, 188 Shizi Street, Jiangsu, Suzhou 215006, China
| | - Liquan Chen
- Department of Critical Care Medicine, The First Affiliated Hospital of Soochow University, 188 Shizi Street, Jiangsu, Suzhou 215006, China
| | - Chang Xu
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, 188 Shizi Street, Jiangsu, Suzhou 215006, China
- Suzhou Clinical Center of Digestive Diseases, Suzhou, China
| | - Shiqi Zhu
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, 188 Shizi Street, Jiangsu, Suzhou 215006, China
- Suzhou Clinical Center of Digestive Diseases, Suzhou, China
| | - Jingwen Gao
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, 188 Shizi Street, Jiangsu, Suzhou 215006, China
- Suzhou Clinical Center of Digestive Diseases, Suzhou, China
| | - Lu Liu
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, 188 Shizi Street, Jiangsu, Suzhou 215006, China
- Suzhou Clinical Center of Digestive Diseases, Suzhou, China
| | - Xiaolin Liu
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, 188 Shizi Street, Jiangsu, Suzhou 215006, China
- Suzhou Clinical Center of Digestive Diseases, Suzhou, China
| | - Chenqi Gu
- Department of Radiology, The First Affiliated Hospital of Soochow University, Suzhou, China
| | - Zhou Huang
- Department of Radiology, The First Affiliated Hospital of Soochow University, Suzhou, China
| | - Yao Wei
- Department of Critical Care Medicine, The First Affiliated Hospital of Soochow University, 188 Shizi Street, Jiangsu, Suzhou 215006, China.
| | - Jinzhou Zhu
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, 188 Shizi Street, Jiangsu, Suzhou 215006, China.
- Suzhou Clinical Center of Digestive Diseases, Suzhou, China.
| |
Collapse
|
14
|
Chen J, Wen Y, Pokojovy M, Tseng TLB, McCaffrey P, Vo A, Walser E, Moen S. Multi-modal learning for inpatient length of stay prediction. Comput Biol Med 2024; 171:108121. [PMID: 38382388 DOI: 10.1016/j.compbiomed.2024.108121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Revised: 12/20/2023] [Accepted: 02/04/2024] [Indexed: 02/23/2024]
Abstract
Predicting inpatient length of stay (LoS) is important for hospitals aiming to improve service efficiency and enhance management capabilities. Patient medical records are strongly associated with LoS. However, due to diverse modalities, heterogeneity, and complexity of data, it becomes challenging to effectively leverage these heterogeneous data to put forth a predictive model that can accurately predict LoS. To address the challenge, this study aims to establish a novel data-fusion model, termed as DF-Mdl, to integrate heterogeneous clinical data for predicting the LoS of inpatients between hospital discharge and admission. Multi-modal data such as demographic data, clinical notes, laboratory test results, and medical images are utilized in our proposed methodology with individual "basic" sub-models separately applied to each different data modality. Specifically, a convolutional neural network (CNN) model, which we termed CRXMDL, is designed for chest X-ray (CXR) image data, two long short-term memory networks are used to extract features from long text data, and a novel attention-embedded 1D convolutional neural network is developed to extract useful information from numerical data. Finally, these basic models are integrated to form a new data-fusion model (DF-Mdl) for inpatient LoS prediction. The proposed method attains the best R2 and EVAR values of 0.6039 and 0.6042 among competitors for the LoS prediction on the Medical Information Mart for Intensive Care (MIMIC)-IV test dataset. Empirical evidence suggests better performance compared with other state-of-the-art (SOTA) methods, which demonstrates the effectiveness and feasibility of the proposed approach.
Collapse
Affiliation(s)
- Junde Chen
- Dale E. and Sarah Ann Fowler School of Engineering, Chapman University, Orange, CA, 92866, USA
| | - Yuxin Wen
- Dale E. and Sarah Ann Fowler School of Engineering, Chapman University, Orange, CA, 92866, USA.
| | - Michael Pokojovy
- Department of Mathematics and Statistics, Old Dominion University, Norfolk, VA, 23529, USA
| | - Tzu-Liang Bill Tseng
- Department of Industrial, Manufacturing and Systems Engineering, The University of Texas at El Paso, El Paso, TX, 79968, USA
| | - Peter McCaffrey
- University of Texas Medical Branch, Galveston, TX, 77550, USA
| | - Alexander Vo
- University of Texas Medical Branch, Galveston, TX, 77550, USA
| | - Eric Walser
- University of Texas Medical Branch, Galveston, TX, 77550, USA
| | - Scott Moen
- University of Texas Medical Branch, Galveston, TX, 77550, USA
| |
Collapse
|
15
|
Tran A, Wang A, Mickaill J, Strbenac D, Larance M, Vernon ST, Grieve SM, Figtree GA, Patrick E, Yang JYH. Construction and optimization of multi-platform precision pathways for precision medicine. Sci Rep 2024; 14:4248. [PMID: 38378802 PMCID: PMC10879206 DOI: 10.1038/s41598-024-54517-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2023] [Accepted: 02/13/2024] [Indexed: 02/22/2024] Open
Abstract
In the enduring challenge against disease, advancements in medical technology have empowered clinicians with novel diagnostic platforms. Whilst in some cases, a single test may provide a confident diagnosis, often additional tests are required. However, to strike a balance between diagnostic accuracy and cost-effectiveness, one must rigorously construct the clinical pathways. Here, we developed a framework to build multi-platform precision pathways in an automated, unbiased way, recommending the key steps a clinician would take to reach a diagnosis. We achieve this by developing a confidence score, used to simulate a clinical scenario, where at each stage, either a confident diagnosis is made, or another test is performed. Our framework provides a range of tools to interpret, visualize and compare the pathways, improving communication and enabling their evaluation on accuracy and cost, specific to different contexts. This framework will guide the development of novel diagnostic pathways for different diseases, accelerating the implementation of precision medicine into clinical practice.
Collapse
Affiliation(s)
- Andy Tran
- School of Mathematics and Statistics, The University of Sydney, Camperdown, NSW, Australia
- Charles Perkins Centre, The University of Sydney, Camperdown, NSW, Australia
- Sydney Precision Data Science Centre, The University of Sydney, Camperdown, NSW, Australia
| | - Andy Wang
- Westmead Medical Institute, Westmead, NSW, Australia
| | - Jamie Mickaill
- School of Mathematics and Statistics, The University of Sydney, Camperdown, NSW, Australia
- School of Computer Science, The University of Sydney, Camperdown, NSW, Australia
| | - Dario Strbenac
- School of Mathematics and Statistics, The University of Sydney, Camperdown, NSW, Australia
- Charles Perkins Centre, The University of Sydney, Camperdown, NSW, Australia
- Sydney Precision Data Science Centre, The University of Sydney, Camperdown, NSW, Australia
| | - Mark Larance
- Charles Perkins Centre, The University of Sydney, Camperdown, NSW, Australia
| | - Stephen T Vernon
- Charles Perkins Centre, The University of Sydney, Camperdown, NSW, Australia
- Kolling Institute of Medical Research, St Leonards, NSW, Australia
| | - Stuart M Grieve
- Charles Perkins Centre, The University of Sydney, Camperdown, NSW, Australia
- Department of Radiology, Royal Prince Alfred Hospital, Camperdown, Australia
| | - Gemma A Figtree
- Charles Perkins Centre, The University of Sydney, Camperdown, NSW, Australia
- Kolling Institute of Medical Research, St Leonards, NSW, Australia
| | - Ellis Patrick
- School of Mathematics and Statistics, The University of Sydney, Camperdown, NSW, Australia
- Charles Perkins Centre, The University of Sydney, Camperdown, NSW, Australia
- Sydney Precision Data Science Centre, The University of Sydney, Camperdown, NSW, Australia
- Laboratory of Data Discovery for Health Limited (D24H), Science Park, Hong Kong SAR, China
| | - Jean Yee Hwa Yang
- School of Mathematics and Statistics, The University of Sydney, Camperdown, NSW, Australia.
- Charles Perkins Centre, The University of Sydney, Camperdown, NSW, Australia.
- Sydney Precision Data Science Centre, The University of Sydney, Camperdown, NSW, Australia.
- Laboratory of Data Discovery for Health Limited (D24H), Science Park, Hong Kong SAR, China.
| |
Collapse
|
16
|
Wenk J, Voigt I, Inojosa H, Schlieter H, Ziemssen T. Building digital patient pathways for the management and treatment of multiple sclerosis. Front Immunol 2024; 15:1356436. [PMID: 38433832 PMCID: PMC10906094 DOI: 10.3389/fimmu.2024.1356436] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2023] [Accepted: 01/30/2024] [Indexed: 03/05/2024] Open
Abstract
Recent advances in the field of artificial intelligence (AI) could yield new insights into the potential causes of multiple sclerosis (MS) and factors influencing its course as the use of AI opens new possibilities regarding the interpretation and use of big data from not only a cross-sectional, but also a longitudinal perspective. For each patient with MS, there is a vast amount of multimodal data being accumulated over time. But for the application of AI and related technologies, these data need to be available in a machine-readable format and need to be collected in a standardized and structured manner. Through the use of mobile electronic devices and the internet it has also become possible to provide healthcare services from remote and collect information on a patient's state of health outside of regular check-ups on site. Against this background, we argue that the concept of pathways in healthcare now could be applied to structure the collection of information across multiple devices and stakeholders in the virtual sphere, enabling us to exploit the full potential of AI technology by e.g., building digital twins. By going digital and using pathways, we can virtually link patients and their caregivers. Stakeholders then could rely on digital pathways for evidence-based guidance in the sequence of procedures and selection of therapy options based on advanced analytics supported by AI as well as for communication and education purposes. As far as we aware of, however, pathway modelling with respect to MS management and treatment has not been thoroughly investigated yet and still needs to be discussed. In this paper, we thus present our ideas for a modular-integrative framework for the development of digital patient pathways for MS treatment.
Collapse
Affiliation(s)
- Judith Wenk
- Center of Clinical Neuroscience, Department of Neurology, Faculty of Medicine and University Hospital Carl Gustav Carus, Technische Universität Dresden, Dresden, Germany
| | - Isabel Voigt
- Center of Clinical Neuroscience, Department of Neurology, Faculty of Medicine and University Hospital Carl Gustav Carus, Technische Universität Dresden, Dresden, Germany
| | - Hernan Inojosa
- Center of Clinical Neuroscience, Department of Neurology, Faculty of Medicine and University Hospital Carl Gustav Carus, Technische Universität Dresden, Dresden, Germany
| | - Hannes Schlieter
- Research Group Digital Health, Faculty of Business and Economics, Technische Universität Dresden, Dresden, Germany
| | - Tjalf Ziemssen
- Center of Clinical Neuroscience, Department of Neurology, Faculty of Medicine and University Hospital Carl Gustav Carus, Technische Universität Dresden, Dresden, Germany
| |
Collapse
|
17
|
Rajendran S, Pan W, Sabuncu MR, Chen Y, Zhou J, Wang F. Learning across diverse biomedical data modalities and cohorts: Challenges and opportunities for innovation. PATTERNS (NEW YORK, N.Y.) 2024; 5:100913. [PMID: 38370129 PMCID: PMC10873158 DOI: 10.1016/j.patter.2023.100913] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 02/20/2024]
Abstract
In healthcare, machine learning (ML) shows significant potential to augment patient care, improve population health, and streamline healthcare workflows. Realizing its full potential is, however, often hampered by concerns about data privacy, diversity in data sources, and suboptimal utilization of different data modalities. This review studies the utility of cross-cohort cross-category (C4) integration in such contexts: the process of combining information from diverse datasets distributed across distinct, secure sites. We argue that C4 approaches could pave the way for ML models that are both holistic and widely applicable. This paper provides a comprehensive overview of C4 in health care, including its present stage, potential opportunities, and associated challenges.
Collapse
Affiliation(s)
- Suraj Rajendran
- Tri-Institutional Computational Biology & Medicine Program, Cornell University, Ithaca, NY, USA
| | - Weishen Pan
- Division of Health Informatics, Department of Population Health Sciences, Weill Cornell Medicine, New York, NY, USA
| | - Mert R. Sabuncu
- School of Electrical and Computer Engineering, Cornell University, Ithaca, NY, USA
- Cornell Tech, Cornell University, New York, NY, USA
- Department of Radiology, Weill Cornell Medical School, New York, NY, USA
| | - Yong Chen
- Department of Biostatistics, Epidemiology, and Informatics, University of Pennsylvania, Philadelphia, PA, USA
| | - Jiayu Zhou
- Department of Computer Science and Engineering, Michigan State University, East Lansing, MI, USA
| | - Fei Wang
- Division of Health Informatics, Department of Population Health Sciences, Weill Cornell Medicine, New York, NY, USA
| |
Collapse
|
18
|
Tong L, Shi W, Isgut M, Zhong Y, Lais P, Gloster L, Sun J, Swain A, Giuste F, Wang MD. Integrating Multi-Omics Data With EHR for Precision Medicine Using Advanced Artificial Intelligence. IEEE Rev Biomed Eng 2024; 17:80-97. [PMID: 37824325 DOI: 10.1109/rbme.2023.3324264] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/14/2023]
Abstract
With the recent advancement of novel biomedical technologies such as high-throughput sequencing and wearable devices, multi-modal biomedical data ranging from multi-omics molecular data to real-time continuous bio-signals are generated at an unprecedented speed and scale every day. For the first time, these multi-modal biomedical data are able to make precision medicine close to a reality. However, due to data volume and the complexity, making good use of these multi-modal biomedical data requires major effort. Researchers and clinicians are actively developing artificial intelligence (AI) approaches for data-driven knowledge discovery and causal inference using a variety of biomedical data modalities. These AI-based approaches have demonstrated promising results in various biomedical and healthcare applications. In this review paper, we summarize the state-of-the-art AI models for integrating multi-omics data and electronic health records (EHRs) for precision medicine. We discuss the challenges and opportunities in integrating multi-omics data with EHRs and future directions. We hope this review can inspire future research and developing in integrating multi-omics data with EHRs for precision medicine.
Collapse
|
19
|
Milosevic M, Jin Q, Singh A, Amal S. Applications of AI in multi-modal imaging for cardiovascular disease. FRONTIERS IN RADIOLOGY 2024; 3:1294068. [PMID: 38283302 PMCID: PMC10811170 DOI: 10.3389/fradi.2023.1294068] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/14/2023] [Accepted: 12/22/2023] [Indexed: 01/30/2024]
Abstract
Data for healthcare is diverse and includes many different modalities. Traditional approaches to Artificial Intelligence for cardiovascular disease were typically limited to single modalities. With the proliferation of diverse datasets and new methods in AI, we are now able to integrate different modalities, such as magnetic resonance scans, computerized tomography scans, echocardiography, x-rays, and electronic health records. In this paper, we review research from the last 5 years in applications of AI to multi-modal imaging. There have been many promising results in registration, segmentation, and fusion of different magnetic resonance imaging modalities with each other and computer tomography scans, but there are still many challenges that need to be addressed. Only a few papers have addressed modalities such as x-ray, echocardiography, or non-imaging modalities. As for prediction or classification tasks, there have only been a couple of papers that use multiple modalities in the cardiovascular domain. Furthermore, no models have been implemented or tested in real world cardiovascular clinical settings.
Collapse
Affiliation(s)
- Marko Milosevic
- Roux Institute, Northeastern University, Portland, ME, United States
| | - Qingchu Jin
- Roux Institute, Northeastern University, Portland, ME, United States
| | - Akarsh Singh
- College of Engineering, Northeastern University, Boston, MA, United States
| | - Saeed Amal
- Roux Institute, Northeastern University, Portland, ME, United States
| |
Collapse
|
20
|
Liu L, Du K. A perspective on computer vision in biosensing. BIOMICROFLUIDICS 2024; 18:011301. [PMID: 38223547 PMCID: PMC10787640 DOI: 10.1063/5.0185732] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/01/2023] [Accepted: 12/26/2023] [Indexed: 01/16/2024]
Abstract
Computer vision has become a powerful tool in the field of biosensing, aiding in the development of innovative and precise systems for the analysis and interpretation of biological data. This interdisciplinary approach harnesses the capabilities of computer vision algorithms and techniques to extract valuable information from various biosensing applications, including medical diagnostics, environmental monitoring, and food health. Despite years of development, there is still significant room for improvement in this area. In this perspective, we outline how computer vision is applied to raw sensor data in biosensors and its advantages to biosensing applications. We then discuss ongoing research and developments in the field and subsequently explore the challenges and opportunities that computer vision faces in biosensor applications. We also suggest directions for future work, ultimately underscoring the significant impact of computer vision on advancing biosensing technologies and their applications.
Collapse
Affiliation(s)
- Li Liu
- Department of Chemical and Environmental Engineering, University of California, Riverside, California 92521, USA
| | - Ke Du
- Department of Chemical and Environmental Engineering, University of California, Riverside, California 92521, USA
| |
Collapse
|
21
|
Schilcher J, Nilsson A, Andlid O, Eklund A. Fusion of electronic health records and radiographic images for a multimodal deep learning prediction model of atypical femur fractures. Comput Biol Med 2024; 168:107704. [PMID: 37980797 DOI: 10.1016/j.compbiomed.2023.107704] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2023] [Revised: 10/15/2023] [Accepted: 11/07/2023] [Indexed: 11/21/2023]
Abstract
Atypical femur fractures (AFF) represent a very rare type of fracture that can be difficult to discriminate radiologically from normal femur fractures (NFF). AFFs are associated with drugs that are administered to prevent osteoporosis-related fragility fractures, which are highly prevalent in the elderly population. Given that these fractures are rare and the radiologic changes are subtle currently only 7% of AFFs are correctly identified, which hinders adequate treatment for most patients with AFF. Deep learning models could be trained to classify automatically a fracture as AFF or NFF, thereby assisting radiologists in detecting these rare fractures. Historically, for this classification task, only imaging data have been used, using convolutional neural networks (CNN) or vision transformers applied to radiographs. However, to mimic situations in which all available data are used to arrive at a diagnosis, we adopted an approach of deep learning that is based on the integration of image data and tabular data (from electronic health records) for 159 patients with AFF and 914 patients with NFF. We hypothesized that the combinatorial data, compiled from all the radiology departments of 72 hospitals in Sweden and the Swedish National Patient Register, would improve classification accuracy, as compared to using only one modality. At the patient level, the area under the ROC curve (AUC) increased from 0.966 to 0.987 when using the integrated set of imaging data and seven pre-selected variables, as compared to only using imaging data. More importantly, the sensitivity increased from 0.796 to 0.903. We found a greater impact of data fusion when only a randomly selected subset of available images was used to make the image and tabular data more balanced for each patient. The AUC then increased from 0.949 to 0.984, and the sensitivity increased from 0.727 to 0.849. These AUC improvements are not large, mainly because of the already excellent performance of the CNN (AUC of 0.966) when only images are used. However, the improvement is clinically highly relevant considering the importance of accuracy in medical diagnostics. We expect an even greater effect when imaging data from a clinical workflow, comprising a more diverse set of diagnostic images, are used.
Collapse
Affiliation(s)
- Jörg Schilcher
- Department of Orthopedics and Experimental and Clinical Medicine, Faculty of Health Science, Linköping University, Linköping, Sweden; Wallenberg Centre for Molecular Medicine, Linköping University, Linköping, Sweden; Center for Medical Image Science and Visualization (CMIV), Linköping University, Linköping, Sweden
| | - Alva Nilsson
- Department of Biomedical Engineering, Linköping University, Linköping, Sweden
| | - Oliver Andlid
- Department of Biomedical Engineering, Linköping University, Linköping, Sweden
| | - Anders Eklund
- Department of Biomedical Engineering, Linköping University, Linköping, Sweden; Division of Statistics and Machine Learning, Department of Computer and Information Science, Linköping University, Linköping, Sweden; Center for Medical Image Science and Visualization (CMIV), Linköping University, Linköping, Sweden.
| |
Collapse
|
22
|
Han S, Kim YB, No JH, Suh DH, Kim K, Ahn S. Predicting Postoperative Hospital Stays Using Nursing Narratives and the Reverse Time Attention (RETAIN) Model: Retrospective Cohort Study. JMIR Med Inform 2023; 11:e45377. [PMID: 38131977 PMCID: PMC10763991 DOI: 10.2196/45377] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/28/2022] [Revised: 08/02/2023] [Accepted: 08/09/2023] [Indexed: 12/23/2023] Open
Abstract
Background Nursing narratives are an intriguing feature in the prediction of short-term clinical outcomes. However, it is unclear which nursing narratives significantly impact the prediction of postoperative length of stay (LOS) in deep learning models. Objective Therefore, we applied the Reverse Time Attention (RETAIN) model to predict LOS, entering nursing narratives as the main input. Methods A total of 354 patients who underwent ovarian cancer surgery at the Seoul National University Bundang Hospital from 2014 to 2020 were retrospectively enrolled. Nursing narratives collected within 3 postoperative days were used to predict prolonged LOS (≥10 days). The physician's assessment was conducted based on a retrospective review of the physician's note within the same period of the data model used. Results The model performed better than the physician's assessment (area under the receiver operating curve of 0.81 vs 0.58; P=.02). Nursing narratives entered on the first day were the most influential predictors in prolonged LOS. The likelihood of prolonged LOS increased if the physician had to check the patient often and if the patient received intravenous fluids or intravenous patient-controlled analgesia late. Conclusions The use of the RETAIN model on nursing narratives predicted postoperative LOS effectively for patients who underwent ovarian cancer surgery. These findings suggest that accurate and interpretable deep learning information obtained shortly after surgery may accurately predict prolonged LOS.
Collapse
Affiliation(s)
- Sungjoo Han
- Division of Statistics, Medical Research Collaborating Center, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
| | - Yong Bum Kim
- Department of Obstetrics and Gynecology, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
| | - Jae Hong No
- Department of Obstetrics and Gynecology, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
| | - Dong Hoon Suh
- Department of Obstetrics and Gynecology, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
| | - Kidong Kim
- Department of Obstetrics and Gynecology, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
| | - Soyeon Ahn
- Division of Statistics, Medical Research Collaborating Center, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
| |
Collapse
|
23
|
Singh A, Randive S, Breggia A, Ahmad B, Christman R, Amal S. Enhancing Prostate Cancer Diagnosis with a Novel Artificial Intelligence-Based Web Application: Synergizing Deep Learning Models, Multimodal Data, and Insights from Usability Study with Pathologists. Cancers (Basel) 2023; 15:5659. [PMID: 38067363 PMCID: PMC10705310 DOI: 10.3390/cancers15235659] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2023] [Revised: 11/22/2023] [Accepted: 11/23/2023] [Indexed: 05/29/2024] Open
Abstract
Prostate cancer remains a significant cause of male cancer mortality in the United States, with an estimated 288,300 new cases in 2023. Accurate grading of prostate cancer is crucial for ascertaining disease severity and shaping treatment strategies. Modern deep learning techniques show promise in grading biopsies, but there is a gap in integrating these advances into clinical practice. Our web platform tackles this challenge by integrating human expertise with AI-driven grading, incorporating diverse data sources. We gathered feedback from four pathologists and one medical practitioner to assess usability and real-world alignment through a survey and the NASA TLX Usability Test. Notably, 60% of users found it easy to navigate, rating it 5.5 out of 7 for ease of understanding. Users appreciated self-explanatory information in popup tabs. For ease of use, all users favored the detailed summary tab, rating it 6.5 out of 7. While 80% felt patient demographics beyond age were unnecessary, high-resolution biopsy images were deemed vital. Acceptability was high, with all users willing to adopt the app, and some believed it could reduce workload. The NASA TLX Usability Test indicated a low-moderate perceived workload, suggesting room for improved explanations and data visualization.
Collapse
Affiliation(s)
- Akarsh Singh
- College of Engineering, Northeastern University, Boston, MA 02115, USA; (A.S.); (S.R.)
| | - Shruti Randive
- College of Engineering, Northeastern University, Boston, MA 02115, USA; (A.S.); (S.R.)
| | - Anne Breggia
- Maine Health Institute for Research, Scarborough, ME 04074, USA
| | - Bilal Ahmad
- Maine Medical Center, Portland, ME 04102, USA; (B.A.); (R.C.)
| | | | - Saeed Amal
- The Roux Institute, Department of Bioengineering, College of Engineering, Northeastern University, Boston, MA 02115, USA
| |
Collapse
|
24
|
Otsuka N, Kawanishi Y, Doi F, Takeda T, Okumura K, Yamauchi T, Yada S, Wakamiya S, Aramaki E, Makinodan M. Diagnosing psychiatric disorders from history of present illness using a large-scale linguistic model. Psychiatry Clin Neurosci 2023; 77:597-604. [PMID: 37526294 DOI: 10.1111/pcn.13580] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/29/2023] [Revised: 07/22/2023] [Accepted: 07/27/2023] [Indexed: 08/02/2023]
Abstract
AIM Recent advances in natural language processing models are expected to provide diagnostic assistance in psychiatry from the history of present illness (HPI). However, existing studies have been limited, with the target diseases including only major diseases, small sample sizes, or no comparison with diagnoses made by psychiatrists to ensure accuracy. Therefore, we formulated an accurate diagnostic model that covers all psychiatric disorders. METHODS HPIs and diagnoses were extracted from discharge summaries of 2,642 cases at the Nara Medical University Hospital, Japan, from 21 May 2007, to 31 May 31 2021. The diagnoses were classified into 11 classes according to the code from ICD-10 Chapter V. Using UTH-BERT pre-trained on the electronic medical records of the University of Tokyo Hospital, Japan, we predicted the main diagnoses at discharge based on HPIs and compared the concordance rate with the results of psychiatrists. The psychiatrists were divided into two groups: semi-Designated with 3-4 years of experience and Residents with only 2 months of experience. RESULTS The model's match rate was 74.3%, compared to 71.5% for the semi-Designated psychiatrists and 69.4% for the Residents. If the cases were limited to those correctly answered by the semi-Designated group, the model and the Residents performed at 84.9% and 83.3%, respectively. CONCLUSION We demonstrated that the model matched the diagnosis predicted from the HPI with a high probability to the principal diagnosis at discharge. Hence, the model can provide diagnostic suggestions in actual clinical practice.
Collapse
Affiliation(s)
- Norio Otsuka
- Department of Psychiatry, Nara Medical University, Kashihara, Japan
| | - Yuu Kawanishi
- Department of Psychiatry, Nara Medical University, Kashihara, Japan
| | - Fumimaro Doi
- Department of Psychiatry, Nara Medical University, Kashihara, Japan
| | - Tsutomu Takeda
- Department of Psychiatry, Nara Medical University, Kashihara, Japan
| | - Kazuki Okumura
- Department of Psychiatry, Nara Medical University, Kashihara, Japan
| | | | - Shuntaro Yada
- Graduate School of Science and Technology, Nara Institute of Science and Technology, Ikoma, Japan
| | - Shoko Wakamiya
- Graduate School of Science and Technology, Nara Institute of Science and Technology, Ikoma, Japan
| | - Eiji Aramaki
- Graduate School of Science and Technology, Nara Institute of Science and Technology, Ikoma, Japan
| | - Manabu Makinodan
- Department of Psychiatry, Nara Medical University, Kashihara, Japan
| |
Collapse
|
25
|
Boussioux L, Ma Y, Thomas NK, Bertsimas D, Shusharina N, Pursley J, Chen YL, DeLaney TF, Qian J, Bortfeld T. Automated Segmentation of Sacral Chordoma and Surrounding Muscles Using Deep Learning Ensemble. Int J Radiat Oncol Biol Phys 2023; 117:738-749. [PMID: 37451472 PMCID: PMC10665084 DOI: 10.1016/j.ijrobp.2023.03.078] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2022] [Revised: 03/18/2023] [Accepted: 03/30/2023] [Indexed: 07/18/2023]
Abstract
PURPOSE The manual segmentation of organ structures in radiation oncology treatment planning is a time-consuming and highly skilled task, particularly when treating rare tumors like sacral chordomas. This study evaluates the performance of automated deep learning (DL) models in accurately segmenting the gross tumor volume (GTV) and surrounding muscle structures of sacral chordomas. METHODS AND MATERIALS An expert radiation oncologist contoured 5 muscle structures (gluteus maximus, gluteus medius, gluteus minimus, paraspinal, piriformis) and sacral chordoma GTV on computed tomography images from 48 patients. We trained 6 DL auto-segmentation models based on 3-dimensional U-Net and residual 3-dimensional U-Net architectures. We then implemented an average and an optimally weighted average ensemble to improve prediction performance. We evaluated algorithms with the average and standard deviation of the volumetric Dice similarity coefficient, surface Dice similarity coefficient with 2- and 3-mm thresholds, and average symmetric surface distance. One independent expert radiation oncologist assessed the clinical viability of the DL contours and determined the necessary amount of editing before they could be used in clinical practice. RESULTS Quantitatively, the ensembles performed the best across all structures. The optimal ensemble (volumetric Dice similarity coefficient, average symmetric surface distance) was (85.5 ± 6.4, 2.6 ± 0.8; GTV), (94.4 ± 1.5, 1.0 ± 0.4; gluteus maximus), (92.6 ± 0.9, 0.9 ± 0.1; gluteus medius), (85.0 ± 2.7, 1.1 ± 0.3; gluteus minimus), (92.1 ± 1.5, 0.8 ± 0.2; paraspinal), and (78.3 ± 5.7, 1.5 ± 0.6; piriformis). The qualitative evaluation suggested that the best model could reduce the total muscle and tumor delineation time to a 19-minute average. CONCLUSIONS Our methodology produces expert-level muscle and sacral chordoma tumor segmentation using DL and ensemble modeling. It can substantially augment the streamlining and accuracy of treatment planning and represents a critical step toward automated delineation of the clinical target volume in sarcoma and other disease sites.
Collapse
Affiliation(s)
- Leonard Boussioux
- Operations Research Center, Massachusetts Institute of Technology, Cambridge, Massachusetts; Sloan School of Management, Massachusetts Institute of Technology, Cambridge, Massachusetts; University of Washington, Michael G. Foster School of Business, Department of Information Systems and Operations Management, Seattle, Washington.
| | - Yu Ma
- Operations Research Center, Massachusetts Institute of Technology, Cambridge, Massachusetts; Sloan School of Management, Massachusetts Institute of Technology, Cambridge, Massachusetts
| | - Nancy Knight Thomas
- Operations Research Center, Massachusetts Institute of Technology, Cambridge, Massachusetts; Sloan School of Management, Massachusetts Institute of Technology, Cambridge, Massachusetts
| | - Dimitris Bertsimas
- Operations Research Center, Massachusetts Institute of Technology, Cambridge, Massachusetts; Sloan School of Management, Massachusetts Institute of Technology, Cambridge, Massachusetts
| | - Nadya Shusharina
- Department of Radiation Oncology, Massachusetts General Hospital, Boston, Massachusetts; Harvard Medical School, Boston, Massachusetts
| | - Jennifer Pursley
- Department of Radiation Oncology, Massachusetts General Hospital, Boston, Massachusetts; Harvard Medical School, Boston, Massachusetts
| | - Yen-Lin Chen
- Department of Radiation Oncology, Massachusetts General Hospital, Boston, Massachusetts; Harvard Medical School, Boston, Massachusetts
| | - Thomas F DeLaney
- Department of Radiation Oncology, Massachusetts General Hospital, Boston, Massachusetts; Harvard Medical School, Boston, Massachusetts
| | - Jack Qian
- Department of Radiation Oncology, Massachusetts General Hospital, Boston, Massachusetts; Harvard Medical School, Boston, Massachusetts
| | - Thomas Bortfeld
- Department of Radiation Oncology, Massachusetts General Hospital, Boston, Massachusetts; Harvard Medical School, Boston, Massachusetts
| |
Collapse
|
26
|
Skorburg JA, Kupferschmidt KL, Taylor GW. "Large Language Models" Do Much More than Just Language: Some Bioethical Implications of Multi-Modal AI. THE AMERICAN JOURNAL OF BIOETHICS : AJOB 2023; 23:110-113. [PMID: 37812107 DOI: 10.1080/15265161.2023.2250318] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/10/2023]
Affiliation(s)
| | | | - Graham W Taylor
- University of Guelph
- Vector Institute for Artificial Intelligence
| |
Collapse
|
27
|
Pellegrini C, Navab N, Kazi A. Unsupervised pre-training of graph transformers on patient population graphs. Med Image Anal 2023; 89:102895. [PMID: 37473609 DOI: 10.1016/j.media.2023.102895] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2022] [Revised: 07/06/2023] [Accepted: 07/07/2023] [Indexed: 07/22/2023]
Abstract
Pre-training has shown success in different areas of machine learning, such as Computer Vision, Natural Language Processing (NLP), and medical imaging. However, it has not been fully explored for clinical data analysis. An immense amount of clinical records are recorded, but still, data and labels can be scarce for data collected in small hospitals or dealing with rare diseases. In such scenarios, pre-training on a larger set of unlabeled clinical data could improve performance. In this paper, we propose novel unsupervised pre-training techniques designed for heterogeneous, multi-modal clinical data for patient outcome prediction inspired by masked language modeling (MLM), by leveraging graph deep learning over population graphs. To this end, we further propose a graph-transformer-based network, designed to handle heterogeneous clinical data. By combining masking-based pre-training with a transformer-based network, we translate the success of masking-based pre-training in other domains to heterogeneous clinical data. We show the benefit of our pre-training method in a self-supervised and a transfer learning setting, utilizing three medical datasets TADPOLE, MIMIC-III, and a Sepsis Prediction Dataset. We find that our proposed pre-training methods help in modeling the data at a patient and population level and improve performance in different fine-tuning tasks on all datasets.
Collapse
Affiliation(s)
- Chantal Pellegrini
- Computer Aided Medical Procedures, Technical University of Munich, Munich, Germany.
| | - Nassir Navab
- Computer Aided Medical Procedures, Technical University of Munich, Munich, Germany; Computer Aided Medical Procedures, Johns Hopkins University, Baltimore, USA
| | - Anees Kazi
- Computer Aided Medical Procedures, Technical University of Munich, Munich, Germany; Massachusetts General Hospital, Harvard Medical School, Cambridge, MA, USA
| |
Collapse
|
28
|
Viana JN, Pilbeam C, Howard M, Scholz B, Ge Z, Fisser C, Mitchell I, Raman S, Leach J. Maintaining High-Touch in High-Tech Digital Health Monitoring and Multi-Omics Prognostication: Ethical, Equity, and Societal Considerations in Precision Health for Palliative Care. OMICS : A JOURNAL OF INTEGRATIVE BIOLOGY 2023; 27:461-473. [PMID: 37861713 DOI: 10.1089/omi.2023.0120] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/21/2023]
Abstract
Advances in digital health, systems biology, environmental monitoring, and artificial intelligence (AI) continue to revolutionize health care, ushering a precision health future. More than disease treatment and prevention, precision health aims at maintaining good health throughout the lifespan. However, how can precision health impact care for people with a terminal or life-limiting condition? We examine here the ethical, equity, and societal/relational implications of two precision health modalities, (1) integrated systems biology/multi-omics analysis for disease prognostication and (2) digital health technologies for health status monitoring and communication. We focus on three main ethical and societal considerations: benefits and risks associated with integration of these modalities into the palliative care system; inclusion of underrepresented and marginalized groups in technology development and deployment; and the impact of high-tech modalities on palliative care's highly personalized and "high-touch" practice. We conclude with 10 recommendations for ensuring that precision health technologies, such as multi-omics prognostication and digital health monitoring, for palliative care are developed, tested, and implemented ethically, inclusively, and equitably.
Collapse
Affiliation(s)
- John Noel Viana
- Australian National Centre for the Public Awareness of Science, College of Science, The Australian National University, Canberra, Australia
- Responsible Innovation Future Science Platform, Commonwealth Scientific and Industrial Research Organisation, Brisbane, Australia
| | - Caitlin Pilbeam
- School of Medicine and Psychology, College of Health and Medicine, The Australian National University, Canberra, Australia
| | - Mark Howard
- Monash Data Futures Institute, Monash University, Clayton, Australia
- Department of Philosophy, School of Philosophical, Historical and International Studies, Monash University, Clayton, Australia
| | - Brett Scholz
- School of Medicine and Psychology, College of Health and Medicine, The Australian National University, Canberra, Australia
| | - Zongyuan Ge
- Monash Data Futures Institute, Monash University, Clayton, Australia
- Department of Data Science & AI, Monash University, Clayton, Australia
| | - Carys Fisser
- Australian National Centre for the Public Awareness of Science, College of Science, The Australian National University, Canberra, Australia
- School of Medicine and Psychology, College of Health and Medicine, The Australian National University, Canberra, Australia
| | - Imogen Mitchell
- School of Medicine and Psychology, College of Health and Medicine, The Australian National University, Canberra, Australia
- Intensive Care Unit, Canberra Hospital, Canberra, Australia
| | - Sujatha Raman
- Australian National Centre for the Public Awareness of Science, College of Science, The Australian National University, Canberra, Australia
| | - Joan Leach
- Australian National Centre for the Public Awareness of Science, College of Science, The Australian National University, Canberra, Australia
| |
Collapse
|
29
|
Liu J, Capurro D, Nguyen A, Verspoor K. Attention-based multimodal fusion with contrast for robust clinical prediction in the face of missing modalities. J Biomed Inform 2023; 145:104466. [PMID: 37549722 DOI: 10.1016/j.jbi.2023.104466] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2023] [Revised: 06/09/2023] [Accepted: 08/01/2023] [Indexed: 08/09/2023]
Abstract
OBJECTIVE With the increasing amount and growing variety of healthcare data, multimodal machine learning supporting integrated modeling of structured and unstructured data is an increasingly important tool for clinical machine learning tasks. However, it is non-trivial to manage the differences in dimensionality, volume, and temporal characteristics of data modalities in the context of a shared target task. Furthermore, patients can have substantial variations in the availability of data, while existing multimodal modeling methods typically assume data completeness and lack a mechanism to handle missing modalities. METHODS We propose a Transformer-based fusion model with modality-specific tokens that summarize the corresponding modalities to achieve effective cross-modal interaction accommodating missing modalities in the clinical context. The model is further refined by inter-modal, inter-sample contrastive learning to improve the representations for better predictive performance. We denote the model as Attention-based cRoss-MOdal fUsion with contRast (ARMOUR). We evaluate ARMOUR using two input modalities (structured measurements and unstructured text), six clinical prediction tasks, and two evaluation regimes, either including or excluding samples with missing modalities. RESULTS Our model shows improved performances over unimodal or multimodal baselines in both evaluation regimes, including or excluding patients with missing modalities in the input. The contrastive learning improves the representation power and is shown to be essential for better results. The simple setup of modality-specific tokens enables ARMOUR to handle patients with missing modalities and allows comparison with existing unimodal benchmark results. CONCLUSION We propose a multimodal model for robust clinical prediction to achieve improved performance while accommodating patients with missing modalities. This work could inspire future research to study the effective incorporation of multiple, more complex modalities of clinical data into a single model.
Collapse
Affiliation(s)
- Jinghui Liu
- Australian e-Health Research Centre, CSIRO, Queensland, Australia; School of Computing and Information Systems, The University of Melbourne, Victoria, Australia
| | - Daniel Capurro
- School of Computing and Information Systems, The University of Melbourne, Victoria, Australia; Centre for Digital Transformation of Health, The University of Melbourne, Victoria, Australia
| | - Anthony Nguyen
- Australian e-Health Research Centre, CSIRO, Queensland, Australia
| | - Karin Verspoor
- School of Computing and Information Systems, The University of Melbourne, Victoria, Australia; School of Computing Technologies, RMIT University, Victoria, Australia.
| |
Collapse
|
30
|
Wornow M, Xu Y, Thapa R, Patel B, Steinberg E, Fleming S, Pfeffer MA, Fries J, Shah NH. The shaky foundations of large language models and foundation models for electronic health records. NPJ Digit Med 2023; 6:135. [PMID: 37516790 PMCID: PMC10387101 DOI: 10.1038/s41746-023-00879-8] [Citation(s) in RCA: 29] [Impact Index Per Article: 29.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2023] [Accepted: 07/13/2023] [Indexed: 07/31/2023] Open
Abstract
The success of foundation models such as ChatGPT and AlphaFold has spurred significant interest in building similar models for electronic medical records (EMRs) to improve patient care and hospital operations. However, recent hype has obscured critical gaps in our understanding of these models' capabilities. In this narrative review, we examine 84 foundation models trained on non-imaging EMR data (i.e., clinical text and/or structured data) and create a taxonomy delineating their architectures, training data, and potential use cases. We find that most models are trained on small, narrowly-scoped clinical datasets (e.g., MIMIC-III) or broad, public biomedical corpora (e.g., PubMed) and are evaluated on tasks that do not provide meaningful insights on their usefulness to health systems. Considering these findings, we propose an improved evaluation framework for measuring the benefits of clinical foundation models that is more closely grounded to metrics that matter in healthcare.
Collapse
Affiliation(s)
- Michael Wornow
- Department of Computer Science, Stanford University, Stanford, CA, USA.
| | - Yizhe Xu
- Center for Biomedical Informatics Research, Stanford University School of Medicine, Stanford, CA, USA
| | - Rahul Thapa
- Center for Biomedical Informatics Research, Stanford University School of Medicine, Stanford, CA, USA
| | - Birju Patel
- Center for Biomedical Informatics Research, Stanford University School of Medicine, Stanford, CA, USA
| | - Ethan Steinberg
- Department of Computer Science, Stanford University, Stanford, CA, USA
| | - Scott Fleming
- Center for Biomedical Informatics Research, Stanford University School of Medicine, Stanford, CA, USA
| | - Michael A Pfeffer
- Center for Biomedical Informatics Research, Stanford University School of Medicine, Stanford, CA, USA
- Technology and Digital Services, Stanford Health Care, Palo Alto, CA, USA
| | - Jason Fries
- Center for Biomedical Informatics Research, Stanford University School of Medicine, Stanford, CA, USA
| | - Nigam H Shah
- Center for Biomedical Informatics Research, Stanford University School of Medicine, Stanford, CA, USA
- Technology and Digital Services, Stanford Health Care, Palo Alto, CA, USA
- Department of Medicine, Stanford University School of Medicine, Stanford, CA, USA
- Clinical Excellence Research Center, Stanford University School of Medicine, Stanford, CA, USA
| |
Collapse
|
31
|
Singh AV, Chandrasekar V, Paudel N, Laux P, Luch A, Gemmati D, Tissato V, Prabhu KS, Uddin S, Dakua SP. Integrative toxicogenomics: Advancing precision medicine and toxicology through artificial intelligence and OMICs technology. Biomed Pharmacother 2023; 163:114784. [PMID: 37121152 DOI: 10.1016/j.biopha.2023.114784] [Citation(s) in RCA: 24] [Impact Index Per Article: 24.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2023] [Revised: 04/15/2023] [Accepted: 04/24/2023] [Indexed: 05/02/2023] Open
Abstract
More information about a person's genetic makeup, drug response, multi-omics response, and genomic response is now available leading to a gradual shift towards personalized treatment. Additionally, the promotion of non-animal testing has fueled the computational toxicogenomics as a pivotal part of the next-gen risk assessment paradigm. Artificial Intelligence (AI) has the potential to provid new ways analyzing the patient data and making predictions about treatment outcomes or toxicity. As personalized medicine and toxicogenomics involve huge data processing, AI can expedite this process by providing powerful data processing, analysis, and interpretation algorithms. AI can process and integrate a multitude of data including genome data, patient records, clinical data and identify patterns to derive predictive models anticipating clinical outcomes and assessing the risk of any personalized medicine approaches. In this article, we have studied the current trends and future perspectives in personalized medicine & toxicology, the role of toxicogenomics in connecting the two fields, and the impact of AI on personalized medicine & toxicology. In this work, we also study the key challenges and limitations in personalized medicine, toxicogenomics, and AI in order to fully realize their potential.
Collapse
Affiliation(s)
- Ajay Vikram Singh
- Department of Chemical and Product Safety, German Federal Institute for Risk Assessment (BfR), 10589 Berlin, Germany
| | | | - Namuna Paudel
- Department of Chemistry, Amrit Campus, Institute of Science and Technology, Tribhuvan University, Lainchaur, Kathmandu 44600 Nepal
| | - Peter Laux
- Department of Chemical and Product Safety, German Federal Institute for Risk Assessment (BfR), 10589 Berlin, Germany
| | - Andreas Luch
- Department of Chemical and Product Safety, German Federal Institute for Risk Assessment (BfR), 10589 Berlin, Germany
| | - Donato Gemmati
- Department of Translational Medicine, University of Ferrara, 44121 Ferrara, Italy; Centre Hemostasis & Thrombosis, University of Ferrara, 44121 Ferrara, Italy; Centre for Gender Medicine, University of Ferrara, 44121 Ferrara, Italy
| | - Veronica Tissato
- Department of Translational Medicine, University of Ferrara, 44121 Ferrara, Italy; Centre Hemostasis & Thrombosis, University of Ferrara, 44121 Ferrara, Italy; Centre for Gender Medicine, University of Ferrara, 44121 Ferrara, Italy
| | - Kirti S Prabhu
- Translational Research Institute, Academic Health System, Hamad Medical Corporation, Doha, Qatar
| | - Shahab Uddin
- Translational Research Institute, Academic Health System, Hamad Medical Corporation, Doha, Qatar
| | | |
Collapse
|
32
|
Lechat B, Loffler KA, Reynolds AC, Naik G, Vakulin A, Jennings G, Escourrou P, McEvoy RD, Adams RJ, Catcheside PG, Eckert DJ. High night-to-night variability in sleep apnea severity is associated with uncontrolled hypertension. NPJ Digit Med 2023; 6:57. [PMID: 36991115 DOI: 10.1038/s41746-023-00801-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2022] [Accepted: 03/10/2023] [Indexed: 03/31/2023] Open
Abstract
Obstructive sleep apnea (OSA) severity can vary markedly from night-to-night. However, the impact of night-to-night variability in OSA severity on key cardiovascular outcomes such as hypertension is unknown. Thus, the primary aim of this study is to determine the effects of night-to-night variability in OSA severity on hypertension likelihood. This study uses in-home monitoring of 15,526 adults with ~180 nights per participant with an under-mattress sleep sensor device, plus ~30 repeat blood pressure measures. OSA severity is defined from the mean estimated apnea-hypopnoea index (AHI) over the ~6-month recording period for each participant. Night-to-night variability in severity is determined from the standard deviation of the estimated AHI across recording nights. Uncontrolled hypertension is defined as mean systolic blood pressure ≥140 mmHg and/or mean diastolic blood pressure ≥90 mmHg. Regression analyses are performed adjusted for age, sex, and body mass index. A total of 12,287 participants (12% female) are included in the analyses. Participants in the highest night-to-night variability quartile within each OSA severity category, have a 50-70% increase in uncontrolled hypertension likelihood versus the lowest variability quartile, independent of OSA severity. This study demonstrates that high night-to-night variability in OSA severity is a predictor of uncontrolled hypertension, independent of OSA severity. These findings have important implications for the identification of which OSA patients are most at risk of cardiovascular harm.
Collapse
Affiliation(s)
- Bastien Lechat
- Adelaide Institute for Sleep Health and FHMRI Sleep Health, College of Medicine and Public Health, Flinders University, Adelaide, Australia.
| | - Kelly A Loffler
- Adelaide Institute for Sleep Health and FHMRI Sleep Health, College of Medicine and Public Health, Flinders University, Adelaide, Australia
| | - Amy C Reynolds
- Adelaide Institute for Sleep Health and FHMRI Sleep Health, College of Medicine and Public Health, Flinders University, Adelaide, Australia
| | - Ganesh Naik
- Adelaide Institute for Sleep Health and FHMRI Sleep Health, College of Medicine and Public Health, Flinders University, Adelaide, Australia
| | - Andrew Vakulin
- Adelaide Institute for Sleep Health and FHMRI Sleep Health, College of Medicine and Public Health, Flinders University, Adelaide, Australia
| | - Garry Jennings
- Baker Heart and Diabetes Research Institute, Melbourne, Australia
| | | | - R Doug McEvoy
- Adelaide Institute for Sleep Health and FHMRI Sleep Health, College of Medicine and Public Health, Flinders University, Adelaide, Australia
| | - Robert J Adams
- Adelaide Institute for Sleep Health and FHMRI Sleep Health, College of Medicine and Public Health, Flinders University, Adelaide, Australia
| | - Peter G Catcheside
- Adelaide Institute for Sleep Health and FHMRI Sleep Health, College of Medicine and Public Health, Flinders University, Adelaide, Australia
| | - Danny J Eckert
- Adelaide Institute for Sleep Health and FHMRI Sleep Health, College of Medicine and Public Health, Flinders University, Adelaide, Australia
| |
Collapse
|
33
|
Doria AS. Applications of artificial intelligence in clinical management, research and health administration: imaging perspectives with a focus on hemophilia. Expert Rev Hematol 2023:1-15. [PMID: 36939638 DOI: 10.1080/17474086.2023.2192474] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/21/2023]
Abstract
INTRODUCTION Joints of persons with hemophilia are frequently affected by repetitive hemarthrosis. In this paper concepts, perks and quirks of the use of artificial intelligence (AI), machine learning (ML) and deep learning are reviewed within clinical and research contexts of hemophilia and other blood-induced disorders' patient care, targeted to the imaging diagnosis of hemophilic joints, under the perspective of different stakeholders (radiologists, hematologists, nurses, physiotherapists, technologists, researchers, managers and patients/caregivers). AREAS COVERED Rubrics that determine the suitability of the utilization of AI in blood-induced disorders' patient care, including diagnosis and follow-up of patients are discussed, focusing on features in which AI can replace or augment the role of radiology in the clinical management and in research of patients. Insights on features in the design and conduct of AI projects in which the human intervention remains critical are provided. EXPERT OPINION The author discusses research concepts in radiogenomics, and challenges of the utilization of AI in different healthcare fields such as patient safety, data sharing and privacy regulations, workforce education and future jobs' shortage. Finally, the author proposes alternatives and potential solutions to mitigate challenges in successfully deploying ML algorithms into clinical practice.
Collapse
Affiliation(s)
- Andrea S Doria
- Department of Diagnostic Imaging, Research Institute, The Hospital for Sick Children, Toronto, ON, Canada.,Research Institute, The Hospital for Sick Children, Toronto, ON, Canada.,Department of Medical Imaging, University of Toronto, Toronto, ON, Canada
| |
Collapse
|
34
|
Wang Y, Han C, Zhang L, Liu J, An Q, Yang F. Millimeter-wave radar object classification using knowledge-assisted neural network. Front Neurosci 2022; 16:1075538. [PMID: 36620441 PMCID: PMC9815772 DOI: 10.3389/fnins.2022.1075538] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2022] [Accepted: 11/28/2022] [Indexed: 12/24/2022] Open
Abstract
To improve the cognition and understanding capabilities of artificial intelligence (AI) technology, it is a tendency to explore the human brain learning processing and integrate brain mechanisms or knowledge into neural networks for inspiration and assistance. This paper concentrates on the application of AI technology in advanced driving assistance system. In this field, millimeter-wave radar is essential for elaborate environment perception due to its robustness to adverse conditions. However, it is still challenging for radar object classification in the complex traffic environment. In this paper, a knowledge-assisted neural network (KANN) is proposed for radar object classification. Inspired by the human brain cognition mechanism and algorithms based on human expertise, two kinds of prior knowledge are injected into the neural network to guide its training and improve its classification accuracy. Specifically, image knowledge provides spatial information about samples. It is integrated into an attention mechanism in the early stage of the network to help reassign attention precisely. In the late stage, object knowledge is combined with the deep features extracted from the network. It contains discriminant semantic information about samples. An attention-based injection method is proposed to adaptively allocate weights to the knowledge and deep features, generating more comprehensive and discriminative features. Experimental results on measured data demonstrate that KANN is superior to current methods and the performance is improved with knowledge assistance.
Collapse
Affiliation(s)
- Yanhua Wang
- Radar Research Laboratory, School of Information and Electronics, Beijing Institute of Technology, Beijing, China,Beijing Institute of Technology Chongqing Innovation Center, Chongqing, China,Advanced Technology Research Institute, Beijing Institute of Technology, Jinan, Shandong, China,Electromagnetic Sensing Research Center of CEMEE State Key Laboratory, School of Information and Electronics, Beijing Institute of Technology, Beijing, China
| | - Chang Han
- Radar Research Laboratory, School of Information and Electronics, Beijing Institute of Technology, Beijing, China,Electromagnetic Sensing Research Center of CEMEE State Key Laboratory, School of Information and Electronics, Beijing Institute of Technology, Beijing, China
| | - Liang Zhang
- Radar Research Laboratory, School of Information and Electronics, Beijing Institute of Technology, Beijing, China,Electromagnetic Sensing Research Center of CEMEE State Key Laboratory, School of Information and Electronics, Beijing Institute of Technology, Beijing, China,*Correspondence: Liang Zhang,
| | - Jianhu Liu
- Beijing Rxbit Electronic Technology Co., Ltd., Beijing, China
| | - Qingru An
- Beijing Rxbit Electronic Technology Co., Ltd., Beijing, China
| | - Fei Yang
- Beijing Institute of Technology Chongqing Innovation Center, Chongqing, China
| |
Collapse
|