1
|
Luo Y, Mao C, Sanchez‐Pinto LN, Ahmad FS, Naidech A, Rasmussen L, Pacheco JA, Schneider D, Mithal LB, Dresden S, Holmes K, Carson M, Shah SJ, Khan S, Clare S, Wunderink RG, Liu H, Walunas T, Cooper L, Yue F, Wehbe F, Fang D, Liebovitz DM, Markl M, Michelson KN, McColley SA, Green M, Starren J, Ackermann RT, D'Aquila RT, Adams J, Lloyd‐Jones D, Chisholm RL, Kho A. Northwestern University resource and education development initiatives to advance collaborative artificial intelligence across the learning health system. Learn Health Syst 2024; 8:e10417. [PMID: 39036530 PMCID: PMC11257059 DOI: 10.1002/lrh2.10417] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Revised: 02/22/2024] [Accepted: 02/26/2024] [Indexed: 07/23/2024] Open
Abstract
Introduction The rapid development of artificial intelligence (AI) in healthcare has exposed the unmet need for growing a multidisciplinary workforce that can collaborate effectively in the learning health systems. Maximizing the synergy among multiple teams is critical for Collaborative AI in Healthcare. Methods We have developed a series of data, tools, and educational resources for cultivating the next generation of multidisciplinary workforce for Collaborative AI in Healthcare. We built bulk-natural language processing pipelines to extract structured information from clinical notes and stored them in common data models. We developed multimodal AI/machine learning (ML) tools and tutorials to enrich the toolbox of the multidisciplinary workforce to analyze multimodal healthcare data. We have created a fertile ground to cross-pollinate clinicians and AI scientists and train the next generation of AI health workforce to collaborate effectively. Results Our work has democratized access to unstructured health information, AI/ML tools and resources for healthcare, and collaborative education resources. From 2017 to 2022, this has enabled studies in multiple clinical specialties resulting in 68 peer-reviewed publications. In 2022, our cross-discipline efforts converged and institutionalized into the Center for Collaborative AI in Healthcare. Conclusions Our Collaborative AI in Healthcare initiatives has created valuable educational and practical resources. They have enabled more clinicians, scientists, and hospital administrators to successfully apply AI methods in their daily research and practice, develop closer collaborations, and advanced the institution-level learning health system.
Collapse
Affiliation(s)
- Yuan Luo
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Division of Health and Biomedical Informatics, Department of Preventive MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Chengsheng Mao
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Division of Health and Biomedical Informatics, Department of Preventive MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Lazaro N. Sanchez‐Pinto
- Division of Health and Biomedical Informatics, Department of Preventive MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Division of Critical Care, Department of PediatricsNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Stanley Manne Children's Research InstituteAnn & Robert H. Lurie Children's Hospital of ChicagoChicagoIllinoisUSA
| | - Faraz S. Ahmad
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Division of Cardiology, Department of MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Andrew Naidech
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Division of Neurocritical Care, Department of NeurologyNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Luke Rasmussen
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Division of Health and Biomedical Informatics, Department of Preventive MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Jennifer A. Pacheco
- Center for Genetic MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Daniel Schneider
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
| | - Leena B. Mithal
- Stanley Manne Children's Research InstituteAnn & Robert H. Lurie Children's Hospital of ChicagoChicagoIllinoisUSA
- Division of Infectious Diseases, Department of PediatricsNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Scott Dresden
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Department of Emergency MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Kristi Holmes
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Division of Health and Biomedical Informatics, Department of Preventive MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Galter Health Sciences LibraryNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Matthew Carson
- Galter Health Sciences LibraryNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Sanjiv J. Shah
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Division of Cardiology, Department of MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Seema Khan
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Department of SurgeryNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Susan Clare
- Department of SurgeryNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Richard G. Wunderink
- Division of Critical Care, Department of PediatricsNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Pulmonary and Critical Care Division, Department of MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Huiping Liu
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Department of PharmacologyNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Division of Hematology and Oncology, Department of MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Theresa Walunas
- Division of Health and Biomedical Informatics, Department of Preventive MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Division of General Internal Medicine, Department of MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Center for Health Information PartnershipsInstitute for Public Health and Medicine, Northwestern UniversityChicagoIllinoisUSA
- Department of Microbiology‐ImmunologyNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Lee Cooper
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Division of Health and Biomedical Informatics, Department of Preventive MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Department of PathologyNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Feng Yue
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Department of PathologyNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Department of Biochemistry and Molecular GeneticsNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Firas Wehbe
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Department of SurgeryNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Deyu Fang
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Department of PathologyNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - David M. Liebovitz
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Division of Health and Biomedical Informatics, Department of Preventive MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Division of General Internal Medicine, Department of MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Center for Health Information PartnershipsInstitute for Public Health and Medicine, Northwestern UniversityChicagoIllinoisUSA
| | - Michael Markl
- Department of RadiologyNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Kelly N. Michelson
- Division of Critical Care, Department of PediatricsNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Stanley Manne Children's Research InstituteAnn & Robert H. Lurie Children's Hospital of ChicagoChicagoIllinoisUSA
- Center for Bioethics and Medical Humanities, Institute for Public Health and MedicineNorthwestern UniversityChicagoIllinoisUSA
| | - Susanna A. McColley
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Stanley Manne Children's Research InstituteAnn & Robert H. Lurie Children's Hospital of ChicagoChicagoIllinoisUSA
- Division of Pulmonary and Sleep Medicine, Department of PediatricsNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Marianne Green
- Division of General Internal Medicine, Department of MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Justin Starren
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Division of Health and Biomedical Informatics, Department of Preventive MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Ronald T. Ackermann
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Division of General Internal Medicine, Department of MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Institute for Public Health and MedicineNorthwestern UniversityChicagoIllinoisUSA
| | - Richard T. D'Aquila
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Division of Infectious Diseases, Department of MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - James Adams
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Department of Emergency MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Donald Lloyd‐Jones
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Division of Epidemiology, Department of Preventive MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
| | - Rex L. Chisholm
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Department of SurgeryNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Center for Health Information PartnershipsInstitute for Public Health and Medicine, Northwestern UniversityChicagoIllinoisUSA
| | - Abel Kho
- Northwestern University Clinical and Translational Sciences InstituteChicagoIllinoisUSA
- Institute for Augmented Intelligence in MedicineNorthwestern UniversityChicagoIllinoisUSA
- Division of Health and Biomedical Informatics, Department of Preventive MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Division of General Internal Medicine, Department of MedicineNorthwestern University Feinberg School of MedicineChicagoIllinoisUSA
- Center for Health Information PartnershipsInstitute for Public Health and Medicine, Northwestern UniversityChicagoIllinoisUSA
| |
Collapse
|
2
|
Rani S, Jain A. Optimizing healthcare system by amalgamation of text processing and deep learning: a systematic review. MULTIMEDIA TOOLS AND APPLICATIONS 2023:1-25. [PMID: 37362695 PMCID: PMC10183315 DOI: 10.1007/s11042-023-15539-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/23/2021] [Revised: 05/18/2022] [Accepted: 04/19/2023] [Indexed: 06/28/2023]
Abstract
The explosion of clinical textual data has drawn the attention of researchers. Owing to the abundance of clinical data, it is becoming difficult for healthcare professionals to take real-time measures. The tools and methods are lacking when compared to the amount of clinical data generated every day. This review aims to survey the text processing pipeline with deep learning methods such as CNN, RNN, LSTM, and GRU in the healthcare domain and discuss various applications such as clinical concept detection and extraction, medically aware dialogue systems, sentiment analysis of drug reviews shared online, clinical trial matching, and pharmacovigilance. In addition, we highlighted the major challenges in deploying text processing with deep learning to clinical textual data and identified the scope of research in this domain. Furthermore, we have discussed various resources that can be used in the future to optimize the healthcare domain by amalgamating text processing and deep learning.
Collapse
Affiliation(s)
- Somiya Rani
- Department of Computer Science and Engineering, NSUT East Campus (erstwhile AIACTR), Affiliated to Guru Gobind Singh Indraprastha University, Delhi, India
| | - Amita Jain
- Department of Computer Science and Engineering, Netaji Subhas University of Technology, Delhi, India
| |
Collapse
|
3
|
Zhang Y, Li X, Yang Y, Wang T. Disease- and Drug-Related Knowledge Extraction for Health Management from Online Health Communities Based on BERT-BiGRU-ATT. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2022; 19:16590. [PMID: 36554472 PMCID: PMC9779596 DOI: 10.3390/ijerph192416590] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/09/2022] [Revised: 12/01/2022] [Accepted: 12/06/2022] [Indexed: 06/17/2023]
Abstract
Knowledge extraction from rich text in online health communities can supplement and improve the existing knowledge base, supporting evidence-based medicine and clinical decision making. The extracted time series health management data of users can help users with similar conditions when managing their health. By annotating four relationships, this study constructed a deep learning model, BERT-BiGRU-ATT, to extract disease-medication relationships. A Chinese-pretrained BERT model was used to generate word embeddings for the question-and-answer data from online health communities in China. In addition, the bidirectional gated recurrent unit, combined with an attention mechanism, was employed to capture sequence context features and then to classify text related to diseases and drugs using a softmax classifier and to obtain the time series data provided by users. By using various word embedding training experiments and comparisons with classical models, the superiority of our model in relation to extraction was verified. Based on the knowledge extraction, the evolution of a user's disease progression was analyzed according to the time series data provided by users to further analyze the evolution of the user's disease progression. BERT word embedding, GRU, and attention mechanisms in our research play major roles in knowledge extraction. The knowledge extraction results obtained are expected to supplement and improve the existing knowledge base, assist doctors' diagnosis, and help users with dynamic lifecycle health management, such as user disease treatment management. In future studies, a co-reference resolution can be introduced to further improve the effect of extracting the relationships among diseases, drugs, and drug effects.
Collapse
Affiliation(s)
- Yanli Zhang
- College of Business Administration, Henan Finance University, Zhengzhou 451464, China
- Business School, Henan University, Kaifeng 475004, China
| | - Xinmiao Li
- School of Information Management and Engineering, Shanghai University of Finance and Economics, Shanghai 200433, China
| | - Yu Yang
- School of Information Management and Engineering, Shanghai University of Finance and Economics, Shanghai 200433, China
- China Banking and Insurance Regulatory Commission Neimengu Office, Hohhot 010019, China
| | - Tao Wang
- College of Business Administration, Henan Finance University, Zhengzhou 451464, China
| |
Collapse
|
4
|
|
5
|
Zhao Y, Yu Y, Wang H, Li Y, Deng Y, Jiang G, Luo Y. Machine Learning in Causal Inference: Application in Pharmacovigilance. Drug Saf 2022; 45:459-476. [PMID: 35579811 PMCID: PMC9114053 DOI: 10.1007/s40264-022-01155-6] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 02/09/2022] [Indexed: 01/28/2023]
Abstract
Monitoring adverse drug events or pharmacovigilance has been promoted by the World Health Organization to assure the safety of medicines through a timely and reliable information exchange regarding drug safety issues. We aim to discuss the application of machine learning methods as well as causal inference paradigms in pharmacovigilance. We first reviewed data sources for pharmacovigilance. Then, we examined traditional causal inference paradigms, their applications in pharmacovigilance, and how machine learning methods and causal inference paradigms were integrated to enhance the performance of traditional causal inference paradigms. Finally, we summarized issues with currently mainstream correlation-based machine learning models and how the machine learning community has tried to address these issues by incorporating causal inference paradigms. Our literature search revealed that most existing data sources and tasks for pharmacovigilance were not designed for causal inference. Additionally, pharmacovigilance was lagging in adopting machine learning-causal inference integrated models. We highlight several currently trending directions or gaps to integrate causal inference with machine learning in pharmacovigilance research. Finally, our literature search revealed that the adoption of causal paradigms can mitigate known issues with machine learning models. We foresee that the pharmacovigilance domain can benefit from the progress in the machine learning field.
Collapse
Affiliation(s)
- Yiqing Zhao
- Department of Preventive Medicine, Feinberg School of Medicine, Northwestern University, 750 N Lake Shore Drive, Room 11-189, Chicago, IL, 60611, USA
| | - Yue Yu
- Department of Artificial Intelligence and Informatics, Mayo Clinic, Rochester, MN, 55902, USA
| | - Hanyin Wang
- Department of Preventive Medicine, Feinberg School of Medicine, Northwestern University, 750 N Lake Shore Drive, Room 11-189, Chicago, IL, 60611, USA
| | - Yikuan Li
- Department of Preventive Medicine, Feinberg School of Medicine, Northwestern University, 750 N Lake Shore Drive, Room 11-189, Chicago, IL, 60611, USA
| | - Yu Deng
- Department of Preventive Medicine, Feinberg School of Medicine, Northwestern University, 750 N Lake Shore Drive, Room 11-189, Chicago, IL, 60611, USA
| | - Guoqian Jiang
- Department of Artificial Intelligence and Informatics, Mayo Clinic, Rochester, MN, 55902, USA
| | - Yuan Luo
- Department of Preventive Medicine, Feinberg School of Medicine, Northwestern University, 750 N Lake Shore Drive, Room 11-189, Chicago, IL, 60611, USA.
| |
Collapse
|
6
|
de Oliveira JM, da Costa CA, Antunes RS. Data structuring of electronic health records: a systematic review. HEALTH AND TECHNOLOGY 2021. [DOI: 10.1007/s12553-021-00607-w] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/10/2023]
|
7
|
Shi J, Ye Y, Zhu D, Su L, Huang Y, Huang J. Automatic segmentation of cardiac magnetic resonance images based on multi-input fusion network. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2021; 209:106323. [PMID: 34365312 DOI: 10.1016/j.cmpb.2021.106323] [Citation(s) in RCA: 18] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/02/2021] [Accepted: 07/25/2021] [Indexed: 06/13/2023]
Abstract
PURPOSE Using computer-assisted means to process a large amount of heart image data in order to speed up the diagnosis efficiency and accuracy of medical doctors has become a research worthy of investigation. METHOD Based on the U-Net model, this paper proposes a multi-input fusion network (MIFNet) model based on multi-scale input and feature fusion, which automatically extracts and fuses features of different input scales to realize the detection of Cardiac Magnetic Resonance Images (CMRI). The MIFNet model is trained and verified on the public data set, and then compared with the segmentation models, namely the Fully Convolutional Network (FCN) and DeepLab v1. RESULTS MIFNet model segmentation of CMRI significantly improved the segmentation accuracy, and the Dice value reached 97.238%. Compared with FCN and DeepLab v1, the average Hausdorff distance (HD) was reduced by 16.425%. The capacity parameter of FCN is 124.86% of MIFNet, DeepLab v1 is 103.22% of MIFNet. CONCLUSION Our proposed MIFNet model reduces the amount of parameters and improves the training speed while ensuring the simultaneous segmentation of overlapping targets. It can help clinicians to more quickly check the patient's CMRI focus area, and thereby improving the efficiency of diagnosis.
Collapse
Affiliation(s)
- Jianshe Shi
- Faculty of Mathematics and Computer Science, Quanzhou Normal University, Quanzhou 362000, China; Department of General Surgery, Huaqiao University Affiliated Strait Hospital, Quanzhou, Fujian 362000, China
| | - Yuguang Ye
- Faculty of Mathematics and Computer Science, Quanzhou Normal University, Quanzhou 362000, China; Key Laboratory of Intelligent Computing and Information Processing, Fujian Province University, Quanzhou 362000, China; Fujian Provincial Key Laboratory of Data Intensive Computing, Quanzhou 362000, China
| | - Daxin Zhu
- Faculty of Mathematics and Computer Science, Quanzhou Normal University, Quanzhou 362000, China; Key Laboratory of Intelligent Computing and Information Processing, Fujian Province University, Quanzhou 362000, China; Fujian Provincial Key Laboratory of Data Intensive Computing, Quanzhou 362000, China
| | - Lianta Su
- Faculty of Mathematics and Computer Science, Quanzhou Normal University, Quanzhou 362000, China; Key Laboratory of Intelligent Computing and Information Processing, Fujian Province University, Quanzhou 362000, China
| | - Yifeng Huang
- Department of Diagnostic Radiology, Huaqiao University Affiliated Strait Hospital, Quanzhou, Fujian 362000, China.
| | - Jianlong Huang
- Faculty of Mathematics and Computer Science, Quanzhou Normal University, Quanzhou 362000, China; Key Laboratory of Intelligent Computing and Information Processing, Fujian Province University, Quanzhou 362000, China; Fujian Provincial Key Laboratory of Data Intensive Computing, Quanzhou 362000, China.
| |
Collapse
|
8
|
Abstract
Electronic health records (EHRs) are becoming a vital source of data for healthcare quality improvement, research, and operations. However, much of the most valuable information contained in EHRs remains buried in unstructured text. The field of clinical text mining has advanced rapidly in recent years, transitioning from rule-based approaches to machine learning and, more recently, deep learning. With new methods come new challenges, however, especially for those new to the field. This review provides an overview of clinical text mining for those who are encountering it for the first time (e.g., physician researchers, operational analytics teams, machine learning scientists from other domains). While not a comprehensive survey, this review describes the state of the art, with a particular focus on new tasks and methods developed over the past few years. It also identifies key barriers between these remarkable technical advances and the practical realities of implementation in health systems and in industry.
Collapse
Affiliation(s)
- Bethany Percha
- Department of Medicine and Department of Genetics and Genomic Sciences, Icahn School of Medicine at Mount Sinai, New York, NY 10025, USA;
| |
Collapse
|
9
|
Mitra A, Rawat BPS, McManus DD, Yu H. Relation Classification for Bleeding Events From Electronic Health Records Using Deep Learning Systems: An Empirical Study. JMIR Med Inform 2021; 9:e27527. [PMID: 34255697 PMCID: PMC8285744 DOI: 10.2196/27527] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2021] [Revised: 03/19/2021] [Accepted: 05/30/2021] [Indexed: 01/05/2023] Open
Abstract
BACKGROUND Accurate detection of bleeding events from electronic health records (EHRs) is crucial for identifying and characterizing different common and serious medical problems. To extract such information from EHRs, it is essential to identify the relations between bleeding events and related clinical entities (eg, bleeding anatomic sites and lab tests). With the advent of natural language processing (NLP) and deep learning (DL)-based techniques, many studies have focused on their applicability for various clinical applications. However, no prior work has utilized DL to extract relations between bleeding events and relevant entities. OBJECTIVE In this study, we aimed to evaluate multiple DL systems on a novel EHR data set for bleeding event-related relation classification. METHODS We first expert annotated a new data set of 1046 deidentified EHR notes for bleeding events and their attributes. On this data set, we evaluated three state-of-the-art DL architectures for the bleeding event relation classification task, namely, convolutional neural network (CNN), attention-guided graph convolutional network (AGGCN), and Bidirectional Encoder Representations from Transformers (BERT). We used three BERT-based models, namely, BERT pretrained on biomedical data (BioBERT), BioBERT pretrained on clinical text (Bio+Clinical BERT), and BioBERT pretrained on EHR notes (EhrBERT). RESULTS Our experiments showed that the BERT-based models significantly outperformed the CNN and AGGCN models. Specifically, BioBERT achieved a macro F1 score of 0.842, outperforming both the AGGCN (macro F1 score, 0.828) and CNN models (macro F1 score, 0.763) by 1.4% (P<.001) and 7.9% (P<.001), respectively. CONCLUSIONS In this comprehensive study, we explored and compared different DL systems to classify relations between bleeding events and other medical concepts. On our corpus, BERT-based models outperformed other DL models for identifying the relations of bleeding-related entities. In addition to pretrained contextualized word representation, BERT-based models benefited from the use of target entity representation over traditional sequence representation.
Collapse
Affiliation(s)
- Avijit Mitra
- College of Information and Computer Sciences, University of Massachusetts Amherst, Amherst, MA, United States
| | - Bhanu Pratap Singh Rawat
- College of Information and Computer Sciences, University of Massachusetts Amherst, Amherst, MA, United States
| | - David D McManus
- Department of Medicine, University of Massachusetts Medical School, Worcester, MA, United States
| | - Hong Yu
- College of Information and Computer Sciences, University of Massachusetts Amherst, Amherst, MA, United States.,Department of Medicine, University of Massachusetts Medical School, Worcester, MA, United States.,Department of Computer Science, University of Massachusetts Lowell, Lowell, MA, United States.,Center for Healthcare Organization and Implementation Research, Bedford Veterans Affairs Medical Center, Bedford, MA, United States
| |
Collapse
|
10
|
Deng L, Chen L, Yang T, Liu M, Li S, Jiang T. Constructing High-Fidelity Phenotype Knowledge Graphs for Infectious Diseases With a Fine-Grained Semantic Information Model: Development and Usability Study. J Med Internet Res 2021; 23:e26892. [PMID: 34128811 PMCID: PMC8277235 DOI: 10.2196/26892] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/02/2021] [Revised: 04/01/2021] [Accepted: 05/06/2021] [Indexed: 01/19/2023] Open
Abstract
BACKGROUND Phenotypes characterize the clinical manifestations of diseases and provide important information for diagnosis. Therefore, the construction of phenotype knowledge graphs for diseases is valuable to the development of artificial intelligence in medicine. However, phenotype knowledge graphs in current knowledge bases such as WikiData and DBpedia are coarse-grained knowledge graphs because they only consider the core concepts of phenotypes while neglecting the details (attributes) associated with these phenotypes. OBJECTIVE To characterize the details of disease phenotypes for clinical guidelines, we proposed a fine-grained semantic information model named PhenoSSU (semantic structured unit of phenotypes). METHODS PhenoSSU is an "entity-attribute-value" model by its very nature, and it aims to capture the full semantic information underlying phenotype descriptions with a series of attributes and values. A total of 193 clinical guidelines for infectious diseases from Wikipedia were selected as the study corpus, and 12 attributes from SNOMED-CT were introduced into the PhenoSSU model based on the co-occurrences of phenotype concepts and attribute values. The expressive power of the PhenoSSU model was evaluated by analyzing whether PhenoSSU instances could capture the full semantics underlying the descriptions of the corresponding phenotypes. To automatically construct fine-grained phenotype knowledge graphs, a hybrid strategy that first recognized phenotype concepts with the MetaMap tool and then predicted the attribute values of phenotypes with machine learning classifiers was developed. RESULTS Fine-grained phenotype knowledge graphs of 193 infectious diseases were manually constructed with the BRAT annotation tool. A total of 4020 PhenoSSU instances were annotated in these knowledge graphs, and 3757 of them (89.5%) were found to be able to capture the full semantics underlying the descriptions of the corresponding phenotypes listed in clinical guidelines. By comparison, other information models, such as the clinical element model and the HL7 fast health care interoperability resource model, could only capture the full semantics underlying 48.4% (2034/4020) and 21.8% (914/4020) of the descriptions of phenotypes listed in clinical guidelines, respectively. The hybrid strategy achieved an F1-score of 0.732 for the subtask of phenotype concept recognition and an average weighted accuracy of 0.776 for the subtask of attribute value prediction. CONCLUSIONS PhenoSSU is an effective information model for the precise representation of phenotype knowledge for clinical guidelines, and machine learning can be used to improve the efficiency of constructing PhenoSSU-based knowledge graphs. Our work will potentially shift the focus of medical knowledge engineering from a coarse-grained level to a more fine-grained level.
Collapse
Affiliation(s)
- Lizong Deng
- Center of Systems Medicine, Institute of Basic Medical Sciences, Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
- Suzhou Institute of Systems Medicine, Suzhou, China
| | - Luming Chen
- Center of Systems Medicine, Institute of Basic Medical Sciences, Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
- Suzhou Institute of Systems Medicine, Suzhou, China
| | - Tao Yang
- Center of Systems Medicine, Institute of Basic Medical Sciences, Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
- Suzhou Institute of Systems Medicine, Suzhou, China
| | - Mi Liu
- Jiangsu Institute of Clinical Immunology, Jiangsu Key Laboratory of Clinical Immunology, The First Affiliated Hospital of Soochow University, Suzhou, China
| | - Shicheng Li
- Center of Systems Medicine, Institute of Basic Medical Sciences, Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
- Suzhou Institute of Systems Medicine, Suzhou, China
| | - Taijiao Jiang
- Center of Systems Medicine, Institute of Basic Medical Sciences, Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
- Suzhou Institute of Systems Medicine, Suzhou, China
- Guangzhou Laboratory, Guangzhou, China
| |
Collapse
|
11
|
Alfattni G, Belousov M, Peek N, Nenadic G. Extracting Drug Names and Associated Attributes From Discharge Summaries: Text Mining Study. JMIR Med Inform 2021; 9:e24678. [PMID: 33949962 PMCID: PMC8135022 DOI: 10.2196/24678] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2020] [Revised: 02/15/2021] [Accepted: 02/20/2021] [Indexed: 11/13/2022] Open
Abstract
Background Drug prescriptions are often recorded in free-text clinical narratives; making this information available in a structured form is important to support many health-related tasks. Although several natural language processing (NLP) methods have been proposed to extract such information, many challenges remain. Objective This study evaluates the feasibility of using NLP and deep learning approaches for extracting and linking drug names and associated attributes identified in clinical free-text notes and presents an extensive error analysis of different methods. This study initiated with the participation in the 2018 National NLP Clinical Challenges (n2c2) shared task on adverse drug events and medication extraction. Methods The proposed system (DrugEx) consists of a named entity recognizer (NER) to identify drugs and associated attributes and a relation extraction (RE) method to identify the relations between them. For NER, we explored deep learning-based approaches (ie, bidirectional long-short term memory with conditional random fields [BiLSTM-CRFs]) with various embeddings (ie, word embedding, character embedding [CE], and semantic-feature embedding) to investigate how different embeddings influence the performance. A rule-based method was implemented for RE and compared with a context-aware long-short term memory (LSTM) model. The methods were trained and evaluated using the 2018 n2c2 shared task data. Results The experiments showed that the best model (BiLSTM-CRFs with pretrained word embeddings [PWE] and CE) achieved lenient micro F-scores of 0.921 for NER, 0.927 for RE, and 0.855 for the end-to-end system. NER, which relies on the pretrained word and semantic embeddings, performed better on most individual entity types, but NER with PWE and CE had the highest classification efficiency among the proposed approaches. Extracting relations using the rule-based method achieved higher accuracy than the context-aware LSTM for most relations. Interestingly, the LSTM model performed notably better in the reason-drug relations, the most challenging relation type. Conclusions The proposed end-to-end system achieved encouraging results and demonstrated the feasibility of using deep learning methods to extract medication information from free-text data.
Collapse
Affiliation(s)
- Ghada Alfattni
- Department of Computer Science, University of Manchester, Manchester, United Kingdom.,Department of Computer Science, Jamoum University College, Umm Al-Qura University, Makkah, Saudi Arabia
| | - Maksim Belousov
- Department of Computer Science, University of Manchester, Manchester, United Kingdom
| | - Niels Peek
- Centre for Health Informatics, Division of Informatics, Imaging and Data Sciences, University of Manchester, Manchester, United Kingdom.,National Institute of Health Research Manchester Biomedical Research Centre, Manchester Academic Health Science Centre, University of Manchester, Manchester, United Kingdom.,The Alan Turing Institute, Manchester, United Kingdom
| | - Goran Nenadic
- Department of Computer Science, University of Manchester, Manchester, United Kingdom.,The Alan Turing Institute, Manchester, United Kingdom
| |
Collapse
|
12
|
Weinzierl MA, Maldonado R, Harabagiu SM. The impact of learning Unified Medical Language System knowledge embeddings in relation extraction from biomedical texts. J Am Med Inform Assoc 2021; 27:1556-1567. [PMID: 33029619 DOI: 10.1093/jamia/ocaa205] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2020] [Revised: 07/23/2020] [Accepted: 08/07/2020] [Indexed: 11/14/2022] Open
Abstract
OBJECTIVE We explored how knowledge embeddings (KEs) learned from the Unified Medical Language System (UMLS) Metathesaurus impact the quality of relation extraction on 2 diverse sets of biomedical texts. MATERIALS AND METHODS Two forms of KEs were learned for concepts and relation types from the UMLS Metathesaurus, namely lexicalized knowledge embeddings (LKEs) and unlexicalized KEs. A knowledge embedding encoder (KEE) enabled learning either LKEs or unlexicalized KEs as well as neural models capable of producing LKEs for mentions of biomedical concepts in texts and relation types that are not encoded in the UMLS Metathesaurus. This allowed us to design the relation extraction with knowledge embeddings (REKE) system, which incorporates either LKEs or unlexicalized KEs produced for relation types of interest and their arguments. RESULTS The incorporation of either LKEs or unlexicalized KE in REKE advances the state of the art in relation extraction on 2 relation extraction datasets: the 2010 i2b2/VA dataset and the 2013 Drug-Drug Interaction Extraction Challenge corpus. Moreover, the impact of LKEs is superior, achieving F1 scores of 78.2 and 82.0, respectively. DISCUSSION REKE not only highlights the importance of incorporating knowledge encoded in the UMLS Metathesaurus in a novel way, through 2 possible forms of KEs, but it also showcases the subtleties of incorporating KEs in relation extraction systems. CONCLUSIONS Incorporating LKEs informed by the UMLS Metathesaurus in a relation extraction system operating on biomedical texts shows significant promise. We present the REKE system, which establishes new state-of-the-art results for relation extraction on 2 datasets when using LKEs.
Collapse
Affiliation(s)
- Maxwell A Weinzierl
- Human Language Technology Research Institute, Department of Computer Science, Erik Jonsson School of Engineering & Computer Science, University of Texas at Dallas, Richardson, Texas, USA
| | - Ramon Maldonado
- Human Language Technology Research Institute, Department of Computer Science, Erik Jonsson School of Engineering & Computer Science, University of Texas at Dallas, Richardson, Texas, USA
| | - Sanda M Harabagiu
- Human Language Technology Research Institute, Department of Computer Science, Erik Jonsson School of Engineering & Computer Science, University of Texas at Dallas, Richardson, Texas, USA
| |
Collapse
|
13
|
Reimer AP, Milinovich A. Using UMLS for electronic health data standardization and database design. J Am Med Inform Assoc 2021; 27:1520-1528. [PMID: 32940707 DOI: 10.1093/jamia/ocaa176] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/10/2020] [Revised: 07/08/2020] [Accepted: 07/21/2020] [Indexed: 02/06/2023] Open
Abstract
OBJECTIVE Patients that undergo medical transfer represent 1 patient population that remains infrequently studied due to challenges in aggregating data across multiple domains and sources that are necessary to capture the entire episode of patient care. To facilitate access to and secondary use of transport patient data, we developed the Transport Data Repository that combines data from 3 separate domains and many sources within our health system. METHODS The repository is a relational database anchored by the Unified Medical Language System unique concept identifiers to integrate, map, and standardize the data into a common data model. Primary data domains included sending and receiving hospital encounters, medical transport record, and custom hospital transport log data. A 4-step mapping process was developed: 1) automatic source code match, 2) exact text match, 3) fuzzy matching, and 4) manual matching. RESULTS 431 090 total mappings were generated in the Transport Data Repository, consisting of 69 010 unique concepts with 77% of the data being mapped automatically. Transport Source Data yielded significantly lower mapping results with only 8% of data entities automatically mapped and a significant amount (43%) remaining unmapped. DISCUSSION The multistep mapping process resulted in a majority of data been automatically mapped. Poor matching of transport medical record data is due to the third-party vendor data being generated and stored in a nonstandardized format. CONCLUSION The multistep mapping process developed and implemented is necessary to normalize electronic health data from multiple domains and sources into a common data model to support secondary use of data.
Collapse
Affiliation(s)
- Andrew P Reimer
- Frances Payne Bolton School of Nursing, Case Western Reserve University, Cleveland, Ohio,USA.,Critical Care Transport, Cleveland Clinic, Cleveland, Ohio,USA
| | - Alex Milinovich
- Department of Quantitative Health Sciences, Cleveland Clinic, Cleveland, Ohio,USA
| |
Collapse
|
14
|
Bai T, Guan H, Wang S, Wang Y, Huang L. Traditional Chinese medicine entity relation extraction based on CNN with segment attention. Neural Comput Appl 2021. [DOI: 10.1007/s00521-021-05897-9] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
15
|
Shi X, Yi Y, Xiong Y, Tang B, Chen Q, Wang X, Ji Z, Zhang Y, Xu H. Extracting entities with attributes in clinical text via joint deep learning. J Am Med Inform Assoc 2021; 26:1584-1591. [PMID: 31550346 DOI: 10.1093/jamia/ocz158] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2019] [Revised: 07/18/2019] [Accepted: 08/15/2019] [Indexed: 11/13/2022] Open
Abstract
OBJECTIVE Extracting clinical entities and their attributes is a fundamental task of natural language processing (NLP) in the medical domain. This task is typically recognized as 2 sequential subtasks in a pipeline, clinical entity or attribute recognition followed by entity-attribute relation extraction. One problem of pipeline methods is that errors from entity recognition are unavoidably passed to relation extraction. We propose a novel joint deep learning method to recognize clinical entities or attributes and extract entity-attribute relations simultaneously. MATERIALS AND METHODS The proposed method integrates 2 state-of-the-art methods for named entity recognition and relation extraction, namely bidirectional long short-term memory with conditional random field and bidirectional long short-term memory, into a unified framework. In this method, relation constraints between clinical entities and attributes and weights of the 2 subtasks are also considered simultaneously. We compare the method with other related methods (ie, pipeline methods and other joint deep learning methods) on an existing English corpus from SemEval-2015 and a newly developed Chinese corpus. RESULTS Our proposed method achieves the best F1 of 74.46% on entity recognition and the best F1 of 50.21% on relation extraction on the English corpus, and 89.32% and 88.13% on the Chinese corpora, respectively, which outperform the other methods on both tasks. CONCLUSIONS The joint deep learning-based method could improve both entity recognition and relation extraction from clinical text in both English and Chinese, indicating that the approach is promising.
Collapse
Affiliation(s)
- Xue Shi
- Department of Computer Science, Harbin Institute of Technology Shenzhen Graduate School, Shenzhen, China
| | - Yingping Yi
- Department of Science and Education, The Second Affiliated Hospital of Nanchang University, Nanchang, China
| | - Ying Xiong
- Department of Computer Science, Harbin Institute of Technology Shenzhen Graduate School, Shenzhen, China
| | - Buzhou Tang
- Department of Computer Science, Harbin Institute of Technology Shenzhen Graduate School, Shenzhen, China.,Peng Cheng Laboratory
| | - Qingcai Chen
- Department of Computer Science, Harbin Institute of Technology Shenzhen Graduate School, Shenzhen, China
| | - Xiaolong Wang
- Department of Computer Science, Harbin Institute of Technology Shenzhen Graduate School, Shenzhen, China
| | - Zongcheng Ji
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, Texas, USA
| | - Yaoyun Zhang
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, Texas, USA
| | - Hua Xu
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, Texas, USA
| |
Collapse
|
16
|
Neural network’s selection of color in UI design of social software. Neural Comput Appl 2021. [DOI: 10.1007/s00521-020-05422-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
17
|
Ye J, Yao L, Shen J, Janarthanam R, Luo Y. Predicting mortality in critically ill patients with diabetes using machine learning and clinical notes. BMC Med Inform Decis Mak 2020; 20:295. [PMID: 33380338 PMCID: PMC7772896 DOI: 10.1186/s12911-020-01318-4] [Citation(s) in RCA: 21] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2020] [Accepted: 11/09/2020] [Indexed: 12/14/2022] Open
Abstract
BACKGROUND Diabetes mellitus is a prevalent metabolic disease characterized by chronic hyperglycemia. The avalanche of healthcare data is accelerating precision and personalized medicine. Artificial intelligence and algorithm-based approaches are becoming more and more vital to support clinical decision-making. These methods are able to augment health care providers by taking away some of their routine work and enabling them to focus on critical issues. However, few studies have used predictive modeling to uncover associations between comorbidities in ICU patients and diabetes. This study aimed to use Unified Medical Language System (UMLS) resources, involving machine learning and natural language processing (NLP) approaches to predict the risk of mortality. METHODS We conducted a secondary analysis of Medical Information Mart for Intensive Care III (MIMIC-III) data. Different machine learning modeling and NLP approaches were applied. Domain knowledge in health care is built on the dictionaries created by experts who defined the clinical terminologies such as medications or clinical symptoms. This knowledge is valuable to identify information from text notes that assert a certain disease. Knowledge-guided models can automatically extract knowledge from clinical notes or biomedical literature that contains conceptual entities and relationships among these various concepts. Mortality classification was based on the combination of knowledge-guided features and rules. UMLS entity embedding and convolutional neural network (CNN) with word embeddings were applied. Concept Unique Identifiers (CUIs) with entity embeddings were utilized to build clinical text representations. RESULTS The best configuration of the employed machine learning models yielded a competitive AUC of 0.97. Machine learning models along with NLP of clinical notes are promising to assist health care providers to predict the risk of mortality of critically ill patients. CONCLUSION UMLS resources and clinical notes are powerful and important tools to predict mortality in diabetic patients in the critical care setting. The knowledge-guided CNN model is effective (AUC = 0.97) for learning hidden features.
Collapse
Affiliation(s)
- Jiancheng Ye
- Feinberg School of Medicine, Northwestern University, Chicago, IL, USA
| | | | - Jiahong Shen
- Dept. of Materials Science and Engineering, Northwestern University, Evanston, IL, USA
| | | | - Yuan Luo
- Feinberg School of Medicine, Northwestern University, Chicago, IL, USA.
| |
Collapse
|
18
|
Annotating social determinants of health using active learning, and characterizing determinants using neural event extraction. J Biomed Inform 2020; 113:103631. [PMID: 33290878 DOI: 10.1016/j.jbi.2020.103631] [Citation(s) in RCA: 35] [Impact Index Per Article: 8.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2020] [Revised: 11/13/2020] [Accepted: 11/18/2020] [Indexed: 01/19/2023]
Abstract
Social determinants of health (SDOH) affect health outcomes, and knowledge of SDOH can inform clinical decision-making. Automatically extracting SDOH information from clinical text requires data-driven information extraction models trained on annotated corpora that are heterogeneous and frequently include critical SDOH. This work presents a new corpus with SDOH annotations, a novel active learning framework, and the first extraction results on the new corpus. The Social History Annotation Corpus (SHAC) includes 4480 social history sections with detailed annotation for 12 SDOH characterizing the status, extent, and temporal information of 18K distinct events. We introduce a novel active learning framework that selects samples for annotation using a surrogate text classification task as a proxy for a more complex event extraction task. The active learning framework successfully increases the frequency of health risk factors and improves automatic extraction of these events over undirected annotation. An event extraction model trained on SHAC achieves high extraction performance for substance use status (0.82-0.93 F1), employment status (0.81-0.86 F1), and living status type (0.81-0.93 F1) on data from three institutions.
Collapse
|
19
|
Barash Y, Guralnik G, Tau N, Soffer S, Levy T, Shimon O, Zimlichman E, Konen E, Klang E. Comparison of deep learning models for natural language processing-based classification of non-English head CT reports. Neuroradiology 2020; 62:1247-1256. [PMID: 32335686 DOI: 10.1007/s00234-020-02420-0] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2020] [Accepted: 03/26/2020] [Indexed: 02/07/2023]
Abstract
PURPOSE Natural language processing (NLP) can be used for automatic flagging of radiology reports. We assessed deep learning models for classifying non-English head CT reports. METHODS We retrospectively collected head CT reports (2011-2018). Reports were signed in Hebrew. Emergency department (ED) reports of adult patients from January to February for each year (2013-2018) were manually labeled. All other reports were used to pre-train an embedding layer. We explored two use cases: (1) general labeling use case, in which reports were labeled as normal vs. pathological; (2) specific labeling use case, in which reports were labeled as with and without intra-cranial hemorrhage. We tested long short-term memory (LSTM) and LSTM-attention (LSTM-ATN) networks for classifying reports. We also evaluated the improvement of adding Word2Vec word embedding. Deep learning models were compared with a bag-of-words (BOW) model. RESULTS We retrieved 176,988 head CT reports for pre-training. We manually labeled 7784 reports as normal (46.3%) or pathological (53.7%), and 7.1% with intra-cranial hemorrhage. For the general labeling, LSTM-ATN-Word2Vec showed the best results (AUC = 0.967 ± 0.006, accuracy 90.8% ± 0.01). For the specific labeling, all methods showed similar accuracies between 95.0 and 95.9%. Both LSTM-ATN-Word2Vec and BOW had the highest AUC (0.970). CONCLUSION For a general use case, word embedding using a large cohort of non-English head CT reports and ATN improves NLP performance. For a more specific task, BOW and deep learning showed similar results. Models should be explored and tailored to the NLP task.
Collapse
Affiliation(s)
- Yiftach Barash
- Division of Diagnostic Imaging, Sheba Medical Center, Sackler Faculty of Medicine, Tel Aviv University, Derech Sheba St 2, Ramat Gan, Israel.,DeepVision Lab, Sheba Medical Center, Ramat Gan, Israel
| | | | - Noam Tau
- Division of Diagnostic Imaging, Sheba Medical Center, Sackler Faculty of Medicine, Tel Aviv University, Derech Sheba St 2, Ramat Gan, Israel
| | - Shelly Soffer
- Division of Diagnostic Imaging, Sheba Medical Center, Sackler Faculty of Medicine, Tel Aviv University, Derech Sheba St 2, Ramat Gan, Israel.,DeepVision Lab, Sheba Medical Center, Ramat Gan, Israel.,Management, Sheba Medical Center, Sackler Faculty of Medicine, Tel Aviv University, Ramat Gan, Israel
| | - Tal Levy
- DeepVision Lab, Sheba Medical Center, Ramat Gan, Israel.,Tel Aviv University, Tel Aviv, Israel
| | | | - Eyal Zimlichman
- Management, Sheba Medical Center, Sackler Faculty of Medicine, Tel Aviv University, Ramat Gan, Israel
| | - Eli Konen
- Division of Diagnostic Imaging, Sheba Medical Center, Sackler Faculty of Medicine, Tel Aviv University, Derech Sheba St 2, Ramat Gan, Israel
| | - Eyal Klang
- Division of Diagnostic Imaging, Sheba Medical Center, Sackler Faculty of Medicine, Tel Aviv University, Derech Sheba St 2, Ramat Gan, Israel. .,DeepVision Lab, Sheba Medical Center, Ramat Gan, Israel.
| |
Collapse
|
20
|
Spasic I, Nenadic G. Clinical Text Data in Machine Learning: Systematic Review. JMIR Med Inform 2020; 8:e17984. [PMID: 32229465 PMCID: PMC7157505 DOI: 10.2196/17984] [Citation(s) in RCA: 115] [Impact Index Per Article: 28.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2020] [Revised: 02/24/2020] [Accepted: 02/24/2020] [Indexed: 12/22/2022] Open
Abstract
Background Clinical narratives represent the main form of communication within health care, providing a personalized account of patient history and assessments, and offering rich information for clinical decision making. Natural language processing (NLP) has repeatedly demonstrated its feasibility to unlock evidence buried in clinical narratives. Machine learning can facilitate rapid development of NLP tools by leveraging large amounts of text data. Objective The main aim of this study was to provide systematic evidence on the properties of text data used to train machine learning approaches to clinical NLP. We also investigated the types of NLP tasks that have been supported by machine learning and how they can be applied in clinical practice. Methods Our methodology was based on the guidelines for performing systematic reviews. In August 2018, we used PubMed, a multifaceted interface, to perform a literature search against MEDLINE. We identified 110 relevant studies and extracted information about text data used to support machine learning, NLP tasks supported, and their clinical applications. The data properties considered included their size, provenance, collection methods, annotation, and any relevant statistics. Results The majority of datasets used to train machine learning models included only hundreds or thousands of documents. Only 10 studies used tens of thousands of documents, with a handful of studies utilizing more. Relatively small datasets were utilized for training even when much larger datasets were available. The main reason for such poor data utilization is the annotation bottleneck faced by supervised machine learning algorithms. Active learning was explored to iteratively sample a subset of data for manual annotation as a strategy for minimizing the annotation effort while maximizing the predictive performance of the model. Supervised learning was successfully used where clinical codes integrated with free-text notes into electronic health records were utilized as class labels. Similarly, distant supervision was used to utilize an existing knowledge base to automatically annotate raw text. Where manual annotation was unavoidable, crowdsourcing was explored, but it remains unsuitable because of the sensitive nature of data considered. Besides the small volume, training data were typically sourced from a small number of institutions, thus offering no hard evidence about the transferability of machine learning models. The majority of studies focused on text classification. Most commonly, the classification results were used to support phenotyping, prognosis, care improvement, resource management, and surveillance. Conclusions We identified the data annotation bottleneck as one of the key obstacles to machine learning approaches in clinical NLP. Active learning and distant supervision were explored as a way of saving the annotation efforts. Future research in this field would benefit from alternatives such as data augmentation and transfer learning, or unsupervised learning, which do not require data annotation.
Collapse
Affiliation(s)
- Irena Spasic
- School of Computer Science and Informatics, Cardiff University, Cardiff, United Kingdom
| | - Goran Nenadic
- Department of Computer Science, University of Manchester, Manchester, United Kingdom
| |
Collapse
|
21
|
Wei Q, Ji Z, Si Y, Du J, Wang J, Tiryaki F, Wu S, Tao C, Roberts K, Xu H. Relation Extraction from Clinical Narratives Using Pre-trained Language Models. AMIA ... ANNUAL SYMPOSIUM PROCEEDINGS. AMIA SYMPOSIUM 2020; 2019:1236-1245. [PMID: 32308921 PMCID: PMC7153059] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Natural language processing (NLP) is useful for extracting information from clinical narratives, and both traditional machine learning methods and more-recent deep learning methods have been successful in various clinical NLP tasks. These methods often depend on traditional word embeddings that are outputs of language models (LMs). Recently, methods that are directly based on pre-trained language models themselves, followed by fine-tuning on the LMs (e.g. the Bidirectional Encoder Representations from Transformers (BERT)), have achieved state-of-the-art performance on many NLP tasks. Despite their success in the open domain and biomedical literature, these pre-trained LMs have not yet been applied to the clinical relation extraction (RE) task. In this study, we developed two different implementations of the BERT model for clinical RE tasks. Our results show that our tuned LMs outperformed previous state-of-the-art RE systems in two shared tasks, which demonstrates the potential of LM-based methods on the RE task.
Collapse
Affiliation(s)
- Qiang Wei
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, USA
| | - Zongcheng Ji
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, USA
| | - Yuqi Si
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, USA
| | - Jingcheng Du
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, USA
| | - Jingqi Wang
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, USA
| | - Firat Tiryaki
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, USA
| | - Stephen Wu
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, USA
| | - Cui Tao
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, USA
| | - Kirk Roberts
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, USA
| | - Hua Xu
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, USA
| |
Collapse
|
22
|
Li Y, Jin R, Luo Y. Classifying relations in clinical narratives using segment graph convolutional and recurrent neural networks (Seg-GCRNs). J Am Med Inform Assoc 2019; 26:262-268. [PMID: 30590613 DOI: 10.1093/jamia/ocy157] [Citation(s) in RCA: 29] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/05/2018] [Accepted: 11/03/2018] [Indexed: 01/16/2023] Open
Abstract
We propose to use segment graph convolutional and recurrent neural networks (Seg-GCRNs), which use only word embedding and sentence syntactic dependencies, to classify relations from clinical notes without manual feature engineering. In this study, the relations between 2 medical concepts are classified by simultaneously learning representations of text segments in the context of sentence syntactic dependency: preceding, concept1, middle, concept2, and succeeding segments. Seg-GCRN was systematically evaluated on the i2b2/VA relation classification challenge datasets. Experiments show that Seg-GCRN attains state-of-the-art micro-averaged F-measure for all 3 relation categories: 0.692 for classifying medical treatment-problem relations, 0.827 for medical test-problem relations, and 0.741 for medical problem-medical problem relations. Comparison with the previous state-of-the-art segment convolutional neural network (Seg-CNN) suggests that adding syntactic dependency information helps refine medical word embedding and improves concept relation classification without manual feature engineering. Seg-GCRN can be trained efficiently for the i2b2/VA dataset on a GPU platform.
Collapse
Affiliation(s)
- Yifu Li
- Grado Department of Industrial and Systems Engineering (ISE), Virginia Tech, Blacksburg, Virginia, USA
| | - Ran Jin
- Grado Department of Industrial and Systems Engineering (ISE), Virginia Tech, Blacksburg, Virginia, USA
| | - Yuan Luo
- Department of Preventive Medicine, Northwestern University, Chicago, Illinois, USA
| |
Collapse
|
23
|
Jackson G, Hu J. Artificial Intelligence in Health in 2018: New Opportunities, Challenges, and Practical Implications. Yearb Med Inform 2019; 28:52-54. [PMID: 31419815 PMCID: PMC6697508 DOI: 10.1055/s-0039-1677925] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/07/2023] Open
Abstract
Objective
: To summarize significant research contributions to the field of artificial intelligence (AI) in health in 2018.
Methods
: Ovid MEDLINE
®
and Web of Science
®
databases were searched to identify original research articles that were published in the English language during 2018 and presented advances in the science of AI applied in health. Queries employed Medical Subject Heading (MeSH
®
) terms and keywords representing AI methodologies and limited results to health applications. Section editors selected 15 best paper candidates that underwent peer review by internationally renowned domain experts. Final best papers were selected by the editorial board of the 2018 International Medical Informatics Association (IMIA) Yearbook.
Results
: Database searches returned 1,480 unique publications. Best papers employed innovative AI techniques that incorporated domain knowledge or explored approaches to support distributed or federated learning. All top-ranked papers incorporated novel approaches to advance the science of AI in health and included rigorous evaluations of their methodologies.
Conclusions
: Performance of state-of-the-art AI machine learning algorithms can be enhanced by approaches that employ a multidisciplinary biomedical informatics pipeline to incorporate domain knowledge and can overcome challenges such as sparse, missing, or inconsistent data. Innovative training heuristics and encryption techniques may support distributed learning with preservation of privacy.
Collapse
Affiliation(s)
- Gretchen Jackson
- IBM Watson Health, Cambridge, Massachusetts, USA.,Vanderbilt University Medical Center, Nashville, Tennessee, USA
| | - Jianying Hu
- IBM Research, Yorktown Heights, New York, USA
| | | |
Collapse
|
24
|
Abstract
INTRODUCTION Artificial intelligence (AI) technologies continue to attract interest from a broad range of disciplines in recent years, including health. The increase in computer hardware and software applications in medicine, as well as digitization of health-related data together fuel progress in the development and use of AI in medicine. This progress provides new opportunities and challenges, as well as directions for the future of AI in health. OBJECTIVE The goals of this survey are to review the current state of AI in health, along with opportunities, challenges, and practical implications. This review highlights recent developments over the past five years and directions for the future. METHODS Publications over the past five years reporting the use of AI in health in clinical and biomedical informatics journals, as well as computer science conferences, were selected according to Google Scholar citations. Publications were then categorized into five different classes, according to the type of data analyzed. RESULTS The major data types identified were multi-omics, clinical, behavioral, environmental and pharmaceutical research and development (R&D) data. The current state of AI related to each data type is described, followed by associated challenges and practical implications that have emerged over the last several years. Opportunities and future directions based on these advances are discussed. CONCLUSION Technologies have enabled the development of AI-assisted approaches to healthcare. However, there remain challenges. Work is currently underway to address multi-modal data integration, balancing quantitative algorithm performance and qualitative model interpretability, protection of model security, federated learning, and model bias.
Collapse
Affiliation(s)
- Fei Wang
- Division of Health Informatics, Department of Healthcare Policy and Research, Weill Cornell Medicine, Cornell University, NY, USA
| | | |
Collapse
|
25
|
Li F, Yu H. An investigation of single-domain and multidomain medication and adverse drug event relation extraction from electronic health record notes using advanced deep learning models. J Am Med Inform Assoc 2019; 26:646-654. [PMID: 30938761 PMCID: PMC6562161 DOI: 10.1093/jamia/ocz018] [Citation(s) in RCA: 16] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2018] [Revised: 01/15/2019] [Accepted: 01/31/2019] [Indexed: 11/13/2022] Open
Abstract
OBJECTIVE We aim to evaluate the effectiveness of advanced deep learning models (eg, capsule network [CapNet], adversarial training [ADV]) for single-domain and multidomain relation extraction from electronic health record (EHR) notes. MATERIALS AND METHODS We built multiple deep learning models with increased complexity, namely a multilayer perceptron (MLP) model and a CapNet model for single-domain relation extraction and fully shared (FS), shared-private (SP), and adversarial training (ADV) modes for multidomain relation extraction. Our models were evaluated in 2 ways: first, we compared our models using our expert-annotated cancer (the MADE1.0 corpus) and cardio corpora; second, we compared our models with the systems in the MADE1.0 and i2b2 challenges. RESULTS Multidomain models outperform single-domain models by 0.7%-1.4% in F1 (t test P < .05), but the results of FS, SP, and ADV modes are mixed. Our results show that the MLP model generally outperforms the CapNet model by 0.1%-1.0% in F1. In the comparisons with other systems, the CapNet model achieves the state-of-the-art result (87.2% in F1) in the cancer corpus and the MLP model generally outperforms MedEx in the cancer, cardiovascular diseases, and i2b2 corpora. CONCLUSIONS Our MLP or CapNet model generally outperforms other state-of-the-art systems in medication and adverse drug event relation extraction. Multidomain models perform better than single-domain models. However, neither the SP nor the ADV mode can always outperform the FS mode significantly. Moreover, the CapNet model is not superior to the MLP model for our corpora.
Collapse
Affiliation(s)
- Fei Li
- Department of Computer Science, University of Massachusetts Lowell, Lowell, Massachusetts, USA
- Center for Healthcare Organization and Implementation Research, Bedford Veterans Affairs Medical Center, Bedford, Massachusetts, USA
| | - Hong Yu
- Department of Computer Science, University of Massachusetts Lowell, Lowell, Massachusetts, USA
- Center for Healthcare Organization and Implementation Research, Bedford Veterans Affairs Medical Center, Bedford, Massachusetts, USA
| |
Collapse
|
26
|
Tran T, Kavuluru R. Distant supervision for treatment relation extraction by leveraging MeSH subheadings. Artif Intell Med 2019; 98:18-26. [PMID: 31521249 PMCID: PMC6748648 DOI: 10.1016/j.artmed.2019.06.002] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2018] [Revised: 06/04/2019] [Accepted: 06/05/2019] [Indexed: 11/26/2022]
Abstract
The growing body of knowledge in biomedicine is too vast for human consumption. Hence there is a need for automated systems able to navigate and distill the emerging wealth of information. One fundamental task to that end is relation extraction, whereby linguistic expressions of semantic relationships between biomedical entities are recognized and extracted. In this study, we propose a novel distant supervision approach for relation extraction of binary treatment relationships such that high quality positive/negative training examples are generated from PubMed abstracts by leveraging associated MeSH subheadings. The quality of generated examples is assessed based on the quality of supervised models they induce; that is, the mean performance of trained models (derived via bootstrapped ensembling) on a gold standard test set is used as a proxy for data quality. We show that our approach is preferable to traditional distant supervision for treatment relations and is closer to human crowd annotations in terms of annotation quality. For treatment relations, our generated training data performs at 81.38%, compared to traditional distant supervision at 64.33% and crowd-sourced annotations at 90.57% on the model-wide PR-AUC metric. We also demonstrate that examples generated using our method can be used to augment crowd-sourced datasets. Augmented models improve over non-augmented models by more than two absolute points on the more established F1 metric. We lastly demonstrate that performance can be further improved by implementing a classification loss that is resistant to label noise.
Collapse
Affiliation(s)
- Tung Tran
- Department of Computer Science, University of Kentucky, Lexington, KY, United States.
| | - Ramakanth Kavuluru
- Department of Computer Science, University of Kentucky, Lexington, KY, United States; Division of Biomedical Informatics, Department of Internal Medicine, University of Kentucky, Lexington, KY, United States.
| |
Collapse
|
27
|
Assale M, Dui LG, Cina A, Seveso A, Cabitza F. The Revival of the Notes Field: Leveraging the Unstructured Content in Electronic Health Records. Front Med (Lausanne) 2019; 6:66. [PMID: 31058150 PMCID: PMC6478793 DOI: 10.3389/fmed.2019.00066] [Citation(s) in RCA: 36] [Impact Index Per Article: 7.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/27/2018] [Accepted: 03/18/2019] [Indexed: 01/01/2023] Open
Abstract
Problem: Clinical practice requires the production of a time- and resource-consuming great amount of notes. They contain relevant information, but their secondary use is almost impossible, due to their unstructured nature. Researchers are trying to address this problems, with traditional and promising novel techniques. Application in real hospital settings seems not to be possible yet, though, both because of relatively small and dirty dataset, and for the lack of language-specific pre-trained models. Aim: Our aim is to demonstrate the potential of the above techniques, but also raise awareness of the still open challenges that the scientific communities of IT and medical practitioners must jointly address to realize the full potential of unstructured content that is daily produced and digitized in hospital settings, both to improve its data quality and leverage the insights from data-driven predictive models. Methods: To this extent, we present a narrative literature review of the most recent and relevant contributions to leverage the application of Natural Language Processing techniques to the free-text content electronic patient records. In particular, we focused on four selected application domains, namely: data quality, information extraction, sentiment analysis and predictive models, and automated patient cohort selection. Then, we will present a few empirical studies that we undertook at a major teaching hospital specializing in musculoskeletal diseases. Results: We provide the reader with some simple and affordable pipelines, which demonstrate the feasibility of reaching literature performance levels with a single institution non-English dataset. In such a way, we bridged literature and real world needs, performing a step further toward the revival of notes fields.
Collapse
Affiliation(s)
- Michela Assale
- K-tree SRL, Pont-Saint-Martin, Italy
- University of Milano-Bicocca, Milan, Italy
| | - Linda Greta Dui
- Politecnico di Milano, Milan, Italy
- Link-Up Datareg, Cinisello Balsamo, Italy
| | - Andrea Cina
- K-tree SRL, Pont-Saint-Martin, Italy
- University of Milano-Bicocca, Milan, Italy
| | - Andrea Seveso
- University of Milano-Bicocca, Milan, Italy
- Link-Up Datareg, Cinisello Balsamo, Italy
| | - Federico Cabitza
- University of Milano-Bicocca, Milan, Italy
- IRCCS Istituto Ortopedico Galeazzi, Milan, Italy
| |
Collapse
|
28
|
Zeng Z, Yao L, Roy A, Li X, Espino S, Clare SE, Khan SA, Luo Y. Identifying Breast Cancer Distant Recurrences from Electronic Health Records Using Machine Learning. JOURNAL OF HEALTHCARE INFORMATICS RESEARCH 2019; 3:283-299. [PMID: 33225204 DOI: 10.1007/s41666-019-00046-3] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Abstract
Accurately identifying distant recurrences in breast cancer from the Electronic Health Records (EHR) is important for both clinical care and secondary analysis. Although multiple applications have been developed for computational phenotyping in breast cancer, distant recurrence identification still relies heavily on manual chart review. In this study, we aim to develop a model that identifies distant recurrences in breast cancer using clinical narratives and structured data from EHR. We applied MetaMap to extract features from clinical narratives and also retrieved structured clinical data from EHR. Using these features, we trained a support vector machine model to identify distant recurrences in breast cancer patients. We trained the model using 1,396 double-annotated subjects and validated the model using 599 double-annotated subjects. In addition, we validated the model on a set of 4,904 single-annotated subjects as a generalization test. In the held-out test and generalization test, we obtained F-measure scores of 0.78 and 0.74, area under curve (AUC) scores of 0.95 and 0.93, respectively. To explore the representation learning utility of deep neural networks, we designed multiple convolutional neural networks and multilayer neural networks to identify distant recurrences. Using the same test set and generalizability test set, we obtained F-measure scores of 0.79 ± 0.02 and 0.74 ± 0.004, AUC scores of 0.95 ± 0.002 and 0.95 ± 0.01, respectively. Our model can accurately and efficiently identify distant recurrences in breast cancer by combining features extracted from unstructured clinical narratives and structured clinical data.
Collapse
Affiliation(s)
- Zexian Zeng
- Preventive Medicine, Northwestern University Feinberg School of Medicine, Chicago, IL, USA
| | - Liang Yao
- Preventive Medicine, Northwestern University Feinberg School of Medicine, Chicago, IL, USA
| | - Ankita Roy
- Surgery, Northwestern University Feinberg School of Medicine, Chicago, IL, USA
| | - Xiaoyu Li
- Social and Behavioral Sciences Harvard T.H. Chan School of Public Health, Boston, MA, USA
| | - Sasa Espino
- Surgery, Northwestern University Feinberg School of Medicine, Chicago, IL, USA
| | - Susan E Clare
- Surgery, Northwestern University Feinberg School of Medicine, Chicago, IL, USA
| | - Seema A Khan
- Surgery, Northwestern University Feinberg School of Medicine, Chicago, IL, USA
| | - Yuan Luo
- Preventive Medicine, Northwestern University Feinberg School of Medicine, Chicago, IL, USA
| |
Collapse
|
29
|
Clinical text classification with rule-based features and knowledge-guided convolutional neural networks. BMC Med Inform Decis Mak 2019; 19:71. [PMID: 30943960 PMCID: PMC6448186 DOI: 10.1186/s12911-019-0781-4] [Citation(s) in RCA: 41] [Impact Index Per Article: 8.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022] Open
Abstract
Background Clinical text classification is an fundamental problem in medical natural language processing. Existing studies have cocnventionally focused on rules or knowledge sources-based feature engineering, but only a limited number of studies have exploited effective representation learning capability of deep learning methods. Methods In this study, we propose a new approach which combines rule-based features and knowledge-guided deep learning models for effective disease classification. Critical Steps of our method include recognizing trigger phrases, predicting classes with very few examples using trigger phrases and training a convolutional neural network (CNN) with word embeddings and Unified Medical Language System (UMLS) entity embeddings. Results We evaluated our method on the 2008 Integrating Informatics with Biology and the Bedside (i2b2) obesity challenge. The results demonstrate that our method outperforms the state-of-the-art methods. Conclusion We showed that CNN model is powerful for learning effective hidden features, and CUIs embeddings are helpful for building clinical text representations. This shows integrating domain knowledge into CNN models is promising.
Collapse
|
30
|
Li Z, Yang Z, Shen C, Xu J, Zhang Y, Xu H. Integrating shortest dependency path and sentence sequence into a deep learning framework for relation extraction in clinical text. BMC Med Inform Decis Mak 2019; 19:22. [PMID: 30700301 PMCID: PMC6354333 DOI: 10.1186/s12911-019-0736-9] [Citation(s) in RCA: 26] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/05/2022] Open
Abstract
Background Extracting relations between important clinical entities is critical but very challenging for natural language processing (NLP) in the medical domain. Researchers have applied deep learning-based approaches to clinical relation extraction; but most of them consider sentence sequence only, without modeling syntactic structures. The aim of this study was to utilize a deep neural network to capture the syntactic features and further improve the performances of relation extraction in clinical notes. Methods We propose a novel neural approach to model shortest dependency path (SDP) between target entities together with the sentence sequence for clinical relation extraction. Our neural network architecture consists of three modules: (1) sentence sequence representation module using bidirectional long short-term memory network (Bi-LSTM) to capture the features in the sentence sequence; (2) SDP representation module implementing the convolutional neural network (CNN) and Bi-LSTM network to capture the syntactic context for target entities using SDP information; and (3) classification module utilizing a fully-connected layer with Softmax function to classify the relation type between target entities. Results Using the 2010 i2b2/VA relation extraction dataset, we compared our approach with other baseline methods. Our experimental results show that the proposed approach achieved significant improvements over comparable existing methods, demonstrating the effectiveness of utilizing syntactic structures in deep learning-based relation extraction. The F-measure of our method reaches 74.34% which is 2.5% higher than the method without using syntactic features. Conclusions We propose a new neural network architecture by modeling SDP along with sentence sequence to extract multi-relations from clinical text. Our experimental results show that the proposed approach significantly improve the performances on clinical notes, demonstrating the effectiveness of syntactic structures in deep learning-based relation extraction.
Collapse
Affiliation(s)
- Zhiheng Li
- School of Computer Science and Technology, Dalian University of Technology, Dalian, 116024, China
| | - Zhihao Yang
- School of Computer Science and Technology, Dalian University of Technology, Dalian, 116024, China
| | - Chen Shen
- School of Computer Science and Technology, Dalian University of Technology, Dalian, 116024, China
| | - Jun Xu
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, 77030, USA
| | - Yaoyun Zhang
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, 77030, USA
| | - Hua Xu
- School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, 77030, USA.
| |
Collapse
|
31
|
Zeng Z, Deng Y, Li X, Naumann T, Luo Y. Natural Language Processing for EHR-Based Computational Phenotyping. IEEE/ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS 2019; 16:139-153. [PMID: 29994486 PMCID: PMC6388621 DOI: 10.1109/tcbb.2018.2849968] [Citation(s) in RCA: 90] [Impact Index Per Article: 18.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
This article reviews recent advances in applying natural language processing (NLP) to Electronic Health Records (EHRs) for computational phenotyping. NLP-based computational phenotyping has numerous applications including diagnosis categorization, novel phenotype discovery, clinical trial screening, pharmacogenomics, drug-drug interaction (DDI), and adverse drug event (ADE) detection, as well as genome-wide and phenome-wide association studies. Significant progress has been made in algorithm development and resource construction for computational phenotyping. Among the surveyed methods, well-designed keyword search and rule-based systems often achieve good performance. However, the construction of keyword and rule lists requires significant manual effort, which is difficult to scale. Supervised machine learning models have been favored because they are capable of acquiring both classification patterns and structures from data. Recently, deep learning and unsupervised learning have received growing attention, with the former favored for its performance and the latter for its ability to find novel phenotypes. Integrating heterogeneous data sources have become increasingly important and have shown promise in improving model performance. Often, better performance is achieved by combining multiple modalities of information. Despite these many advances, challenges and opportunities remain for NLP-based computational phenotyping, including better model interpretability and generalizability, and proper characterization of feature relations in clinical narratives.
Collapse
Affiliation(s)
- Zexian Zeng
- Department of Preventive Medicine, Northwestern University Feinberg School of Medicine, Chicago, IL 60611.
| | - Yu Deng
- Department of Preventive Medicine, Northwestern University Feinberg School of Medicine, Chicago, IL 60611.
| | - Xiaoyu Li
- Department of Social and Behavioral Sciences, Harvard T.H. Chan School of Public Health, Boston, MA 02115.
| | - Tristan Naumann
- Science and Artificial Intelligence Lab, Massachusetts Institue of Technology, Cambridge, MA 02139.
| | - Yuan Luo
- Department of Preventive Medicine, Northwestern University Feinberg School of Medicine, Chicago, IL 60611.
| |
Collapse
|
32
|
Chen T, Wu M, Li H. A general approach for improving deep learning-based medical relation extraction using a pre-trained model and fine-tuning. DATABASE-THE JOURNAL OF BIOLOGICAL DATABASES AND CURATION 2019; 2019:5645655. [PMID: 31800044 PMCID: PMC6892305 DOI: 10.1093/database/baz116] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/24/2019] [Revised: 07/16/2019] [Accepted: 09/02/2019] [Indexed: 01/07/2023]
Abstract
The automatic extraction of meaningful relations from biomedical literature or clinical records is crucial in various biomedical applications. Most of the current deep learning approaches for medical relation extraction require large-scale training data to prevent overfitting of the training model. We propose using a pre-trained model and a fine-tuning technique to improve these approaches without additional time-consuming human labeling. Firstly, we show the architecture of Bidirectional Encoder Representations from Transformers (BERT), an approach for pre-training a model on large-scale unstructured text. We then combine BERT with a one-dimensional convolutional neural network (1d-CNN) to fine-tune the pre-trained model for relation extraction. Extensive experiments on three datasets, namely the BioCreative V chemical disease relation corpus, traditional Chinese medicine literature corpus and i2b2 2012 temporal relation challenge corpus, show that the proposed approach achieves state-of-the-art results (giving a relative improvement of 22.2, 7.77, and 38.5% in F1 score, respectively, compared with a traditional 1d-CNN classifier). The source code is available at https://github.com/chentao1999/MedicalRelationExtraction.
Collapse
Affiliation(s)
- Tao Chen
- Department of Computer Science and Engineering, Faculty of Intelligent Manufacturing, Wuyi University, No.22, Dongcheng village, Pengjiang district, Jiangmen City, Guangdong Province, 529020, China
| | - Mingfen Wu
- Department of Computer Science and Engineering, Faculty of Intelligent Manufacturing, Wuyi University, No.22, Dongcheng village, Pengjiang district, Jiangmen City, Guangdong Province, 529020, China
| | - Hexi Li
- Department of Computer Science and Engineering, Faculty of Intelligent Manufacturing, Wuyi University, No.22, Dongcheng village, Pengjiang district, Jiangmen City, Guangdong Province, 529020, China
| |
Collapse
|
33
|
Si Y, Roberts K. A Frame-Based NLP System for Cancer-Related Information Extraction. AMIA ... ANNUAL SYMPOSIUM PROCEEDINGS. AMIA SYMPOSIUM 2018; 2018:1524-1533. [PMID: 30815198 PMCID: PMC6371330] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
We propose a frame-based natural language processing (NLP) method that extracts cancer-related information from clinical narratives. We focus on three frames: cancer diagnosis, cancer therapeutic procedure, and tumor description. We utilize a deep learning-based approach, bidirectional Long Short-term Memory (LSTM) Conditional Random Field (CRF), which uses both character and word embeddings. The system consists of two constituent sequence classifiers: a frame identification (lexical unit) classifier and a frame element classifier. The classifier achieves an F1 of 93.70 for cancer diagnosis, 96.33 for therapeutic procedure, and 87.18 for tumor description. These represent improvements of 10.72, 0.85, and 8.04 over a baseline heuristic, respectively. Additionally, we demonstrate that the combination of both GloVe and MIMIC-III embeddings has the best representational effect. Overall, this study demonstrates the effectiveness of deep learning methods to extract frame semantic information from clinical narratives.
Collapse
Affiliation(s)
- Yuqi Si
- School of Biomedical Informatics The University of Texas Health Science Center at Houston Houston, TX, USA
| | - Kirk Roberts
- School of Biomedical Informatics The University of Texas Health Science Center at Houston Houston, TX, USA
| |
Collapse
|
34
|
Mao C, Pan Y, Zeng Z, Yao L, Luo Y. Deep Generative Classifiers for Thoracic Disease Diagnosis with Chest X-ray Images. PROCEEDINGS. IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE 2018; 2018:1209-1214. [PMID: 31341701 PMCID: PMC6651749 DOI: 10.1109/bibm.2018.8621107] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Thoracic diseases are very serious health problems that plague a large number of people. Chest X-ray is currently one of the most popular methods to diagnose thoracic diseases, playing an important role in the healthcare workflow. However, reading the chest X-ray images and giving an accurate diagnosis remain challenging tasks for expert radiologists. With the success of deep learning in computer vision, a growing number of deep neural network architectures were applied to chest X-ray image classification. However, most of the previous deep neural network classifiers were based on deterministic architectures which are usually very noise-sensitive and are likely to aggravate the overfitting issue. In this paper, to make a deep architecture more robust to noise and to reduce overfitting, we propose using deep generative classifiers to automatically diagnose thorax diseases from the chest X-ray images. Unlike the traditional deterministic classifier, a deep generative classifier has a distribution middle layer in the deep neural network. A sampling layer then draws a random sample from the distribution layer and input it to the following layer for classification. The classifier is generative because the class label is generated from samples of a related distribution. Through training the model with a certain amount of randomness, the deep generative classifiers are expected to be robust to noise and can reduce overfitting and then achieve good performances. We implemented our deep generative classifiers based on a number of well-known deterministic neural network architectures, and tested our models on the chest X-ray14 dataset. The results demonstrated the superiority of deep generative classifiers compared with the corresponding deep deterministic classifiers.
Collapse
Affiliation(s)
- Chengsheng Mao
- Dept. of Preventive Medicine, Northwestern University, Chicago, IL, USA
| | - Yiheng Pan
- Dept. of EECS, Northwestern University , Chicago, IL, USA
| | - Zexian Zeng
- Dept. of Preventive Medicine, Northwestern University, Chicago, IL, USA
| | - Liang Yao
- Dept. of Preventive Medicine, Northwestern University, Chicago, IL, USA
| | - Yuan Luo
- Dept. of Preventive Medicine, Northwestern University, Chicago, IL, USA
| |
Collapse
|
35
|
Luo Y, Szolovits P. Implementing a Portable Clinical NLP System with a Common Data Model - a Lisp Perspective. PROCEEDINGS. IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE 2018; 2018:461-466. [PMID: 33376623 PMCID: PMC7769694 DOI: 10.1109/bibm.2018.8621521] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
This paper presents a Lisp architecture for a portable NLP system, termed LAPNLP, for processing clinical notes. LAPNLP integrates multiple standard, customized and in-house developed NLP tools. Our system facilitates portability across different institutions and data systems by incorporating an enriched Common Data Model (CDM) to standardize necessary data elements. It utilizes UMLS to perform domain adaptation when integrating generic domain NLP tools. It also features stand-off annotations that are specified by positional reference to the original document. We built an interval tree based search engine to efficiently query and retrieve the stand-off annotations by specifying positional requirements. We also developed a utility to convert an inline annotation format to stand-off annotations to enable the reuse of clinical text datasets with in-line annotations. We experimented with our system on several NLP facilitated tasks including computational phenotyping for lymphoma patients and semantic relation extraction for clinical notes. These experiments showcased the broader applicability and utility of LAPNLP.
Collapse
Affiliation(s)
- Yuan Luo
- Dept. of Preventive Medicine, Northwestern University, Chicago, USA
| | | |
Collapse
|
36
|
Li F, Liu W, Yu H. Extraction of Information Related to Adverse Drug Events from Electronic Health Record Notes: Design of an End-to-End Model Based on Deep Learning. JMIR Med Inform 2018; 6:e12159. [PMID: 30478023 PMCID: PMC6288593 DOI: 10.2196/12159] [Citation(s) in RCA: 34] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2018] [Revised: 10/31/2018] [Accepted: 11/09/2018] [Indexed: 12/26/2022] Open
Abstract
BACKGROUND Pharmacovigilance and drug-safety surveillance are crucial for monitoring adverse drug events (ADEs), but the main ADE-reporting systems such as Food and Drug Administration Adverse Event Reporting System face challenges such as underreporting. Therefore, as complementary surveillance, data on ADEs are extracted from electronic health record (EHR) notes via natural language processing (NLP). As NLP develops, many up-to-date machine-learning techniques are introduced in this field, such as deep learning and multi-task learning (MTL). However, only a few studies have focused on employing such techniques to extract ADEs. OBJECTIVE We aimed to design a deep learning model for extracting ADEs and related information such as medications and indications. Since extraction of ADE-related information includes two steps-named entity recognition and relation extraction-our second objective was to improve the deep learning model using multi-task learning between the two steps. METHODS We employed the dataset from the Medication, Indication and Adverse Drug Events (MADE) 1.0 challenge to train and test our models. This dataset consists of 1089 EHR notes of cancer patients and includes 9 entity types such as Medication, Indication, and ADE and 7 types of relations between these entities. To extract information from the dataset, we proposed a deep-learning model that uses a bidirectional long short-term memory (BiLSTM) conditional random field network to recognize entities and a BiLSTM-Attention network to extract relations. To further improve the deep-learning model, we employed three typical MTL methods, namely, hard parameter sharing, parameter regularization, and task relation learning, to build three MTL models, called HardMTL, RegMTL, and LearnMTL, respectively. RESULTS Since extraction of ADE-related information is a two-step task, the result of the second step (ie, relation extraction) was used to compare all models. We used microaveraged precision, recall, and F1 as evaluation metrics. Our deep learning model achieved state-of-the-art results (F1=65.9%), which is significantly higher than that (F1=61.7%) of the best system in the MADE1.0 challenge. HardMTL further improved the F1 by 0.8%, boosting the F1 to 66.7%, whereas RegMTL and LearnMTL failed to boost the performance. CONCLUSIONS Deep learning models can significantly improve the performance of ADE-related information extraction. MTL may be effective for named entity recognition and relation extraction, but it depends on the methods, data, and other factors. Our results can facilitate research on ADE detection, NLP, and machine learning.
Collapse
Affiliation(s)
- Fei Li
- Department of Computer Science, University of Massachusetts Lowell, Lowell, MA, United States
- Center for Healthcare Organization and Implementation Research, Bedford Veterans Affairs Medical Center, Bedford, MA, United States
- Department of Medicine, University of Massachusetts Medical School, Worcester, MA, United States
| | - Weisong Liu
- Department of Computer Science, University of Massachusetts Lowell, Lowell, MA, United States
- Center for Healthcare Organization and Implementation Research, Bedford Veterans Affairs Medical Center, Bedford, MA, United States
- Department of Medicine, University of Massachusetts Medical School, Worcester, MA, United States
| | - Hong Yu
- Department of Computer Science, University of Massachusetts Lowell, Lowell, MA, United States
- Center for Healthcare Organization and Implementation Research, Bedford Veterans Affairs Medical Center, Bedford, MA, United States
- Department of Medicine, University of Massachusetts Medical School, Worcester, MA, United States
- School of Computer Science, University of Massachusetts, Amherst, MA, United States
| |
Collapse
|
37
|
Xiao C, Choi E, Sun J. Opportunities and challenges in developing deep learning models using electronic health records data: a systematic review. J Am Med Inform Assoc 2018; 25:1419-1428. [PMID: 29893864 PMCID: PMC6188527 DOI: 10.1093/jamia/ocy068] [Citation(s) in RCA: 283] [Impact Index Per Article: 47.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2018] [Revised: 05/01/2018] [Accepted: 05/08/2018] [Indexed: 12/14/2022] Open
Abstract
Objective To conduct a systematic review of deep learning models for electronic health record (EHR) data, and illustrate various deep learning architectures for analyzing different data sources and their target applications. We also highlight ongoing research and identify open challenges in building deep learning models of EHRs. Design/method We searched PubMed and Google Scholar for papers on deep learning studies using EHR data published between January 1, 2010, and January 31, 2018. We summarize them according to these axes: types of analytics tasks, types of deep learning model architectures, special challenges arising from health data and tasks and their potential solutions, as well as evaluation strategies. Results We surveyed and analyzed multiple aspects of the 98 articles we found and identified the following analytics tasks: disease detection/classification, sequential prediction of clinical events, concept embedding, data augmentation, and EHR data privacy. We then studied how deep architectures were applied to these tasks. We also discussed some special challenges arising from modeling EHR data and reviewed a few popular approaches. Finally, we summarized how performance evaluations were conducted for each task. Discussion Despite the early success in using deep learning for health analytics applications, there still exist a number of issues to be addressed. We discuss them in detail including data and label availability, the interpretability and transparency of the model, and ease of deployment.
Collapse
Affiliation(s)
- Cao Xiao
- AI for Healthcare, IBM Research, Cambridge, Massachusetts, USA
| | - Edward Choi
- School of Computational Science and Engineering, Georgia Institute of Technology, Atlanta, Georgia, USA
| | - Jimeng Sun
- School of Computational Science and Engineering, Georgia Institute of Technology, Atlanta, Georgia, USA
| |
Collapse
|
38
|
Wong J, Horwitz MM, Zhou L, Toh S. Using machine learning to identify health outcomes from electronic health record data. CURR EPIDEMIOL REP 2018; 5:331-342. [PMID: 30555773 DOI: 10.1007/s40471-018-0165-9] [Citation(s) in RCA: 36] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/18/2022]
Abstract
Purpose of review Electronic health records (EHRs) contain valuable data for identifying health outcomes, but these data also present numerous challenges when creating computable phenotyping algorithms. Machine learning methods could help with some of these challenges. In this review, we discuss four common scenarios that researchers may find helpful for thinking critically about when and for what tasks machine learning may be used to identify health outcomes from EHR data. Recent findings We first consider the conditions in which machine learning may be especially useful with respect to two dimensions of a health outcome: 1) the characteristics of its diagnostic criteria, and 2) the format in which its diagnostic data are usually stored within EHR systems. In the first dimension, we propose that for health outcomes with diagnostic criteria involving many clinical factors, vague definitions, or subjective interpretations, machine learning may be useful for modeling the complex diagnostic decision-making process from a vector of clinical inputs to identify individuals with the health outcome. In the second dimension, we propose that for health outcomes where diagnostic information is largely stored in unstructured formats such as free text or images, machine learning may be useful for extracting and structuring this information as part of a natural language processing system or an image recognition task. We then consider these two dimensions jointly to define four common scenarios of health outcomes. For each scenario, we discuss the potential uses for machine learning - first assuming accurate and complete EHR data and then relaxing these assumptions to accommodate the limitations of real-world EHR systems. We illustrate these four scenarios using concrete examples and describe how recent studies have used machine learning to identify these health outcomes from EHR data. Summary Machine learning has great potential to improve the accuracy and efficiency of health outcome identification from EHR systems, especially under certain conditions. To promote the use of machine learning in EHR-based phenotyping tasks, future work should prioritize efforts to increase the transportability of machine learning algorithms for use in multi-site settings.
Collapse
Affiliation(s)
- Jenna Wong
- Department of Population Medicine, Harvard Medical School and Harvard Pilgrim Health Care Institute, Boston, MA
| | - Mara Murray Horwitz
- Department of Population Medicine, Harvard Medical School and Harvard Pilgrim Health Care Institute, Boston, MA
| | - Li Zhou
- Division of General Internal Medicine and Primary Care, Brigham and Women's Hospital, Boston, MA.,Harvard Medical School, Boston, MA
| | - Sengwee Toh
- Department of Population Medicine, Harvard Medical School and Harvard Pilgrim Health Care Institute, Boston, MA
| |
Collapse
|
39
|
Kim H, Han H. Computer-Aided Multi-Target Management of Emergent Alzheimer's Disease. Bioinformation 2018; 14:167-180. [PMID: 29983487 PMCID: PMC6016757 DOI: 10.6026/97320630014167] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2018] [Revised: 04/29/2018] [Accepted: 04/30/2018] [Indexed: 12/13/2022] Open
Abstract
Alzheimer's disease (AD) represents an enormous global health burden in terms of human suffering and economic cost. AD management requires a shift from the prevailing paradigm targeting pathogenesis to design and develop effective drugs with adequate success in clinical trials. Therefore, it is of interest to report a review on amyloid beta (Aβ) effects and other multi-targets including cholinesterase, NFTs, tau protein and TNF associated with brain cell death to be neuro-protective from AD. It should be noted that these molecules have been generated either by target-based or phenotypic methods. Hence, the use of recent advancements in nanomedicine and other natural compounds screening tools as a feasible alternative for circumventing specific liabilities is realized. We review recent developments in the design and identification of neuro-degenerative compounds against AD generated using current advancements in computational multi-target modeling algorithms reflected by theragnosis (combination of diagnostic tests and therapy) concern.
Collapse
Affiliation(s)
- Hyunjo Kim
- Department of Medical Informatics, Ajou Medical University Hospital, Suwon, Kyeounggido province, South Korea
| | - Hyunwook Han
- Department of Informatics, School of Medicine, CHA University, Seongnam, South Korea
- Institute of Basic Medical Sciences, School of Medicine, CHA University, Seongnam, South Korea
| |
Collapse
|
40
|
Tran T, Kavuluru R. An end-to-end deep learning architecture for extracting protein-protein interactions affected by genetic mutations. DATABASE-THE JOURNAL OF BIOLOGICAL DATABASES AND CURATION 2018; 2018:1-13. [PMID: 30239680 PMCID: PMC6146129 DOI: 10.1093/database/bay092] [Citation(s) in RCA: 21] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/28/2018] [Accepted: 08/13/2018] [Indexed: 11/25/2022]
Abstract
The BioCreative VI Track IV (mining protein interactions and mutations for precision medicine) challenge was organized in 2017 with the goal of applying biomedical text mining methods to support advancements in precision medicine approaches. As part of the challenge, a new dataset was introduced for the purpose of building a supervised relation extraction model capable of taking a test article and returning a list of interacting protein pairs identified by their Entrez Gene IDs. Specifically, such pairs represent proteins participating in a binary protein–protein interaction relation where the interaction is additionally affected by a genetic mutation—referred to as a PPIm relation. In this study, we explore an end-to-end approach for PPIm relation extraction by deploying a three-component pipeline involving deep learning-based named-entity recognition and relation classification models along with a knowledge-based approach for gene normalization. We propose several recall-focused improvements to our original challenge entry that placed second when matching on Entrez Gene ID (exact matching) and on HomoloGene ID. On exact matching, the improved system achieved new competitive test results of 37.78% micro-F1 with a precision of 38.22% and recall of 37.34% that corresponds to an improvement from the prior best system by approximately three micro-F1 points. When matching on HomoloGene IDs, we report similarly competitive test results at 46.17% micro-F1 with a precision and recall of 46.67 and 45.59%, respectively, corresponding to an improvement of more than eight micro-F1 points over the prior best result. The code for our deep learning system is made publicly available at https://github.com/bionlproc/biocppi_extraction.
Collapse
Affiliation(s)
- Tung Tran
- Department of Computer Science, University of Kentucky, Lexington, KY, USA
| | - Ramakanth Kavuluru
- Department of Computer Science, University of Kentucky, Lexington, KY, USA.,Division of Biomedical Informatics, Department of Internal Medicine, University of Kentucky, Lexington, KY, USA
| |
Collapse
|