1
|
Zhang W, Zhao L, Gou H, Gong Y, Zhou Y, Feng Q. PRSCS-Net: Progressive 3D/2D rigid Registration network with the guidance of Single-view Cycle Synthesis. Med Image Anal 2024; 97:103283. [PMID: 39094463 DOI: 10.1016/j.media.2024.103283] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/22/2024] [Revised: 07/08/2024] [Accepted: 07/17/2024] [Indexed: 08/04/2024]
Abstract
The 3D/2D registration for 3D pre-operative images (computed tomography, CT) and 2D intra-operative images (X-ray) plays an important role in image-guided spine surgeries. Conventional iterative-based approaches suffer from time-consuming processes. Existing learning-based approaches require high computational costs and face poor performance on large misalignment because of projection-induced losses or ill-posed reconstruction. In this paper, we propose a Progressive 3D/2D rigid Registration network with the guidance of Single-view Cycle Synthesis, named PRSCS-Net. Specifically, we first introduce the differentiable backward/forward projection operator into the single-view cycle synthesis network, which reconstructs corresponding 3D geometry features from two 2D intra-operative view images (one from the input, and the other from the synthesis). In this way, the problem of limited views during reconstruction can be solved. Subsequently, we employ a self-reconstruction path to extract latent representation from pre-operative 3D CT images. The following pose estimation process will be performed in the 3D geometry feature space, which can solve the dimensional gap, greatly reduce the computational complexity, and ensure that the features extracted from pre-operative and intra-operative images are as relevant as possible to pose estimation. Furthermore, to enhance the ability of our model for handling large misalignment, we develop a progressive registration path, including two sub-registration networks, aiming to estimate the pose parameters via two-step warping volume features. Finally, our proposed method has been evaluated on a public dataset CTSpine1k and an in-house dataset C-ArmLSpine for 3D/2D registration. Results demonstrate that PRSCS-Net achieves state-of-the-art registration performance in terms of registration accuracy, robustness, and generalizability compared with existing methods. Thus, PRSCS-Net has potential for clinical spinal disease surgical planning and surgical navigation systems.
Collapse
Affiliation(s)
- Wencong Zhang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Lei Zhao
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Hang Gou
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Yanggang Gong
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Yujia Zhou
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China.
| | - Qianjin Feng
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China.
| |
Collapse
|
2
|
Lee D, Choi A, Mun JH. Deep Learning-Based Fine-Tuning Approach of Coarse Registration for Ear-Nose-Throat (ENT) Surgical Navigation Systems. Bioengineering (Basel) 2024; 11:941. [PMID: 39329683 PMCID: PMC11428421 DOI: 10.3390/bioengineering11090941] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2024] [Revised: 09/12/2024] [Accepted: 09/17/2024] [Indexed: 09/28/2024] Open
Abstract
Accurate registration between medical images and patient anatomy is crucial for surgical navigation systems in minimally invasive surgeries. This study introduces a novel deep learning-based refinement step to enhance the accuracy of surface registration without disrupting established workflows. The proposed method integrates a machine learning model between conventional coarse registration and ICP fine registration. A deep-learning model was trained using simulated anatomical landmarks with introduced localization errors. The model architecture features global feature-based learning, an iterative prediction structure, and independent processing of rotational and translational components. Validation with silicon-masked head phantoms and CT imaging compared the proposed method to both conventional registration and a recent deep-learning approach. The results demonstrated significant improvements in target registration error (TRE) across different facial regions and depths. The average TRE for the proposed method (1.58 ± 0.52 mm) was significantly lower than that of the conventional (2.37 ± 1.14 mm) and previous deep-learning (2.29 ± 0.95 mm) approaches (p < 0.01). The method showed a consistent performance across various facial regions and enhanced registration accuracy for deeper areas. This advancement could significantly enhance precision and safety in minimally invasive surgical procedures.
Collapse
Affiliation(s)
- Dongjun Lee
- Department of Biomechatronic Engineering, College of Biotechnology and Bioengineering, Sungkyunkwan University, Suwon 16419, Republic of Korea
| | - Ahnryul Choi
- Department of Biomedical Engineering, College of Medicine, Chungbuk National Univeristy, Cheongju 28644, Republic of Korea
| | - Joung Hwan Mun
- Department of Biomechatronic Engineering, College of Biotechnology and Bioengineering, Sungkyunkwan University, Suwon 16419, Republic of Korea
| |
Collapse
|
3
|
Li C, Zhang G, Zhao B, Xie D, Du H, Duan X, Hu Y, Zhang L. Advances of surgical robotics: image-guided classification and application. Natl Sci Rev 2024; 11:nwae186. [PMID: 39144738 PMCID: PMC11321255 DOI: 10.1093/nsr/nwae186] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Revised: 04/19/2024] [Accepted: 05/07/2024] [Indexed: 08/16/2024] Open
Abstract
Surgical robotics application in the field of minimally invasive surgery has developed rapidly and has been attracting increasingly more research attention in recent years. A common consensus has been reached that surgical procedures are to become less traumatic and with the implementation of more intelligence and higher autonomy, which is a serious challenge faced by the environmental sensing capabilities of robotic systems. One of the main sources of environmental information for robots are images, which are the basis of robot vision. In this review article, we divide clinical image into direct and indirect based on the object of information acquisition, and into continuous, intermittent continuous, and discontinuous according to the target-tracking frequency. The characteristics and applications of the existing surgical robots in each category are introduced based on these two dimensions. Our purpose in conducting this review was to analyze, summarize, and discuss the current evidence on the general rules on the application of image technologies for medical purposes. Our analysis gives insight and provides guidance conducive to the development of more advanced surgical robotics systems in the future.
Collapse
Affiliation(s)
- Changsheng Li
- School of Mechatronical Engineering, Beijing Institute of Technology, Beijing 100081, China
| | - Gongzi Zhang
- Department of Orthopedics, Chinese PLA General Hospital, Beijing 100141, China
| | - Baoliang Zhao
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
| | - Dongsheng Xie
- School of Mechatronical Engineering, Beijing Institute of Technology, Beijing 100081, China
- School of Medical Technology, Beijing Institute of Technology, Beijing 100081, China
| | - Hailong Du
- Department of Orthopedics, Chinese PLA General Hospital, Beijing 100141, China
| | - Xingguang Duan
- School of Mechatronical Engineering, Beijing Institute of Technology, Beijing 100081, China
- School of Medical Technology, Beijing Institute of Technology, Beijing 100081, China
| | - Ying Hu
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
| | - Lihai Zhang
- Department of Orthopedics, Chinese PLA General Hospital, Beijing 100141, China
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
| |
Collapse
|
4
|
Sun C, Tong F, Luo J, Wang Y, Ou M, Wu Y, Qiu M, Wu W, Gong Y, Luo Z, Qiao L. A Rapid Head Organ Localization System Based on Clinically Realistic Images: A 3D Two Step Progressive Registration Method with CVH Anatomical Knowledge Mapping. Bioengineering (Basel) 2024; 11:891. [PMID: 39329633 PMCID: PMC11428975 DOI: 10.3390/bioengineering11090891] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/24/2024] [Revised: 08/21/2024] [Accepted: 08/29/2024] [Indexed: 09/28/2024] Open
Abstract
Rapid localization of ROI (Region of Interest) for tomographic medical images (TMIs) is an important foundation for efficient image reading, computer-aided education, and well-informed rights of patients. However, due to the multimodality of clinical TMIs, the complexity of anatomy, and the deformation of organs caused by diseases, it is difficult to have a universal and low-cost method for ROI organ localization. This article focuses on actual concerns of TMIs from medical students, engineers, interdisciplinary researchers, and patients, exploring a universal registration method between the clinical CT/MRI dataset and CVH (Chinese Visible Human) to locate the organ ROI in a low-cost and lightweight way. The proposed method is called Two-step Progressive Registration (TSPR), where the first registration adopts "eye-nose triangle" features to determine the spatial orientation, and the second registration adopts the circular contour to determine the spatial scale, ultimately achieving CVH anatomical knowledge automated mapping. Through experimentation with representative clinical TMIs, the registration results are capable of labeling the ROI in the images well and can adapt to the deformation problem of ROI, as well as local extremum problems that are prone to occur in inter-subject registration. Unlike the ideal requirements for TMIs' data quality in laboratory research, TSPR has good adaptability to incomplete and non-thin-layer quality in real clinical data in a low-cost and lightweight way. This helps medical students, engineers, and interdisciplinary researchers independently browse images, receive computer-aided education, and provide patients with better access to well-informed services, highlighting the potential of digital public health and medical education.
Collapse
Affiliation(s)
- Changjin Sun
- Department of Medical Image, College of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing 400038, China
| | - Fei Tong
- Army Medical Center of PLA, Army Medical University, Chongqing 400010, China
| | - Junjie Luo
- Department of Medical Image, College of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing 400038, China
| | - Yuting Wang
- Department of Medical Image, College of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing 400038, China
| | - Mingwen Ou
- Department of Medical Image, College of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing 400038, China
| | - Yi Wu
- Department of Digital Medicine, College of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing 400038, China
| | - Mingguo Qiu
- Department of Medical Image, College of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing 400038, China
| | - Wenjing Wu
- Department of Radiology, Southwest Hospital, Army Medical University, Chongqing 400038, China
| | - Yan Gong
- Department of Medical Image, College of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing 400038, China
| | - Zhongwen Luo
- Department of Medical Image, College of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing 400038, China
| | - Liang Qiao
- Department of Medical Image, College of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing 400038, China
| |
Collapse
|
5
|
Darzi F, Bocklitz T. A Review of Medical Image Registration for Different Modalities. Bioengineering (Basel) 2024; 11:786. [PMID: 39199744 PMCID: PMC11351674 DOI: 10.3390/bioengineering11080786] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2024] [Revised: 07/26/2024] [Accepted: 07/30/2024] [Indexed: 09/01/2024] Open
Abstract
Medical image registration has become pivotal in recent years with the integration of various imaging modalities like X-ray, ultrasound, MRI, and CT scans, enabling comprehensive analysis and diagnosis of biological structures. This paper provides a comprehensive review of registration techniques for medical images, with an in-depth focus on 2D-2D image registration methods. While 3D registration is briefly touched upon, the primary emphasis remains on 2D techniques and their applications. This review covers registration techniques for diverse modalities, including unimodal, multimodal, interpatient, and intra-patient. The paper explores the challenges encountered in medical image registration, including geometric distortion, differences in image properties, outliers, and optimization convergence, and discusses their impact on registration accuracy and reliability. Strategies for addressing these challenges are highlighted, emphasizing the need for continual innovation and refinement of techniques to enhance the accuracy and reliability of medical image registration systems. The paper concludes by emphasizing the importance of accurate medical image registration in improving diagnosis.
Collapse
Affiliation(s)
- Fatemehzahra Darzi
- Institute of Physical Chemistry, Friedrich Schiller University Jena, Helmholtzweg 4, 07743 Jena, Germany;
| | - Thomas Bocklitz
- Institute of Physical Chemistry, Friedrich Schiller University Jena, Helmholtzweg 4, 07743 Jena, Germany;
- Department of Photonic Data Science, Leibniz Institute of Photonic Technology, Albert-Einstein-Straße 9, 07745 Jena, Germany
| |
Collapse
|
6
|
Ringel MJ, Heiselman JS, Richey WL, Meszoely IM, Jarnagin WR, Miga MI. Comparing regularized Kelvinlet functions and the finite element method for registration of medical images to sparse organ data. Med Image Anal 2024; 96:103221. [PMID: 38824864 DOI: 10.1016/j.media.2024.103221] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2023] [Revised: 05/06/2024] [Accepted: 05/25/2024] [Indexed: 06/04/2024]
Abstract
Image-guided surgery collocates patient-specific data with the physical environment to facilitate surgical decision making. Unfortunately, these guidance systems commonly become compromised by intraoperative soft-tissue deformations. Nonrigid image-to-physical registration methods have been proposed to compensate for deformations, but clinical utility requires compatibility of these techniques with data sparsity and temporal constraints in the operating room. While finite element models can be effective in sparse data scenarios, computation time remains a limitation to widespread deployment. This paper proposes a registration algorithm that uses regularized Kelvinlets, which are analytical solutions to linear elasticity in an infinite domain, to overcome these barriers. This algorithm is demonstrated and compared to finite element-based registration on two datasets: a phantom liver deformation dataset and an in vivo breast deformation dataset. The regularized Kelvinlets algorithm resulted in a significant reduction in computation time compared to the finite element method. Accuracy as evaluated by target registration error was comparable between methods. Average target registration errors were 4.6 ± 1.0 and 3.2 ± 0.8 mm on the liver dataset and 5.4 ± 1.4 and 6.4 ± 1.5 mm on the breast dataset for the regularized Kelvinlets and finite element method, respectively. Limitations of regularized Kelvinlets include the lack of organ-specific geometry and the assumptions of linear elasticity and infinitesimal strain. Despite limitations, this work demonstrates the generalizability of regularized Kelvinlets registration on two soft-tissue elastic organs. This method may improve and accelerate registration for image-guided surgery, and it shows the potential of using regularized Kelvinlets on medical imaging data.
Collapse
Affiliation(s)
- Morgan J Ringel
- Vanderbilt University, Department of Biomedical Engineering, Nashville, TN, USA; Vanderbilt Institute for Surgery and Engineering, Nashville, TN, USA.
| | - Jon S Heiselman
- Vanderbilt University, Department of Biomedical Engineering, Nashville, TN, USA; Vanderbilt Institute for Surgery and Engineering, Nashville, TN, USA; Memorial Sloan-Kettering Cancer Center, Department of Surgery, New York, NY, USA
| | - Winona L Richey
- Vanderbilt University, Department of Biomedical Engineering, Nashville, TN, USA; Vanderbilt Institute for Surgery and Engineering, Nashville, TN, USA
| | - Ingrid M Meszoely
- Vanderbilt University Medical Center, Division of Surgical Oncology, Nashville, TN, USA
| | - William R Jarnagin
- Memorial Sloan-Kettering Cancer Center, Department of Surgery, New York, NY, USA
| | - Michael I Miga
- Vanderbilt University, Department of Biomedical Engineering, Nashville, TN, USA; Vanderbilt Institute for Surgery and Engineering, Nashville, TN, USA
| |
Collapse
|
7
|
Arya GC, Khalid M, Mehla S, Jakhmola V. A review of synthetic strategy, SAR, docking, simulation studies, and mechanism of action of isoxazole derivatives as anticancer agents. J Biomol Struct Dyn 2024; 42:4909-4935. [PMID: 37315986 DOI: 10.1080/07391102.2023.2220819] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2023] [Accepted: 05/29/2023] [Indexed: 06/16/2023]
Abstract
Breast cancer (BC) is a global health concern and the leading cause of cancerous death among women across the world, BC has been characterized by fresh lump in the breast or underarm (armpit), thickened or swollen. Worldwide estimated 9.6 million deaths in 2018-2019. Numerous drugs have been approved by FDA for BC treatment but showed numerous adverse effects like bioavailability issues, selectivity issues, and toxicity issues. Therefore, there is an immediate need to develop new molecules that are non-toxic and more efficient for treating cancer. Isoxazole derivatives have gained popularity over the few years due to their effective antitumor potential. These derivatives work against cancer by inhibiting the thymidylate enzyme, inducing apoptosis, inhibiting tubulin polymerization, protein kinase inhibition, and aromatase inhibition. In this study, we have concentrated on the isoxazole derivative with structure-activity relationship study, various synthesis techniques, mechanism of action, docking, and simulation studies pertaining to BC receptors. Hence the development of isoxazole derivatives with improved therapeutic efficacy will inspire further progress in improving human health.Communicated by Ramaswamy H. Sarma.
Collapse
Affiliation(s)
- Girish Chandra Arya
- University Institute of Pharmaceutical Sciences (UIPS), Chandigarh University, Mohali, India
| | - Mohammad Khalid
- Department of Pharmacognosy, College of Pharmacy, Prince Sattam Bin Abdulaziz University, Al-Kharj, Saudi Arabia
| | - Shefali Mehla
- University Institute of Pharmaceutical Sciences (UIPS), Chandigarh University, Mohali, India
| | - Vikash Jakhmola
- Uttaranchal Institute of Pharmaceutical Sciences, Uttaranchal University, Dehradun, India
| |
Collapse
|
8
|
Ragguett RM, Eagleson R, de Ribaupierre S. Evaluating normalized registration and preprocessing methodologies for the analysis of brain MRI in pediatric patients with shunt-treated hydrocephalus. Front Neurosci 2024; 18:1405363. [PMID: 38887369 PMCID: PMC11182356 DOI: 10.3389/fnins.2024.1405363] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2024] [Accepted: 05/06/2024] [Indexed: 06/20/2024] Open
Abstract
Introduction Registration to a standardized template (i.e. "normalization") is a critical step when performing neuroimaging studies. We present a comparative study involving the evaluation of general-purpose registration algorithms for pediatric patients with shunt treated hydrocephalus. Our sample dataset presents a number of intersecting challenges for registration, representing the potentially large deformations to both brain structures and overall brain shape, artifacts from shunts, and morphological differences corresponding to age. The current study assesses the normalization accuracy of shunt-treated hydrocephalus patients using freely available neuroimaging registration tools. Methods Anatomical neuroimages from eight pediatric patients with shunt-treated hydrocephalus were normalized. Four non-linear registration algorithms were assessed in addition to the preprocessing steps of skull-stripping and bias-correction. Registration accuracy was assessed using the Dice Coefficient (DC) and Hausdorff Distance (HD) in subcortical and cortical regions. Results A total of 592 registrations were performed. On average, normalizations performed using the brain extracted and bias-corrected images had a higher DC and lower HD compared to full head/ non-biased corrected images. The most accurate registration was achieved using SyN by ANTs with skull-stripped and bias corrected images. Without preprocessing, the DARTEL Toolbox was able to produce normalized images with comparable accuracy. The use of a pediatric template as an intermediate registration did not improve normalization. Discussion Using structural neuroimages from patients with shunt-treated pediatric hydrocephalus, it was demonstrated that there are tools which perform well after specified pre-processing steps were taken. Overall, these results provide insight to the performance of registration programs that can be used for normalization of brains with complex pathologies.
Collapse
Affiliation(s)
| | - Roy Eagleson
- School of Biomedical Engineering, Western University, London, ON, Canada
- Department of Electrical and Computer Engineering, Western University, London, ON, Canada
- Centre for Brain and Mind, Western University, London, ON, Canada
| | - Sandrine de Ribaupierre
- School of Biomedical Engineering, Western University, London, ON, Canada
- Centre for Brain and Mind, Western University, London, ON, Canada
- Department of Clinical Neurological Sciences, Schulich School of Medicine, Western University, London, ON, Canada
| |
Collapse
|
9
|
Lambert T, Brunner C, Kil D, Wuyts R, D'Hondt E, Montaldo G, Urban A. A deep learning classification task for brain navigation in rodents using micro-Doppler ultrasound imaging. Heliyon 2024; 10:e27432. [PMID: 38495198 PMCID: PMC10943389 DOI: 10.1016/j.heliyon.2024.e27432] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2023] [Revised: 02/28/2024] [Accepted: 02/28/2024] [Indexed: 03/19/2024] Open
Abstract
Positioning and navigation are essential components of neuroimaging as they improve the quality and reliability of data acquisition, leading to advances in diagnosis, treatment outcomes, and fundamental understanding of the brain. Functional ultrasound imaging is an emerging technology providing high-resolution images of the brain vasculature, allowing for the monitoring of brain activity. However, as the technology is relatively new, there is no standardized tool for inferring the position in the brain from the vascular images. In this study, we present a deep learning-based framework designed to address this challenge. Our approach uses an image classification task coupled with a regression on the resulting probabilities to determine the position of a single image. To evaluate its performance, we conducted experiments using a dataset of 51 rat brain scans. The training positions were extracted at intervals of 375 μm, resulting in a positioning error of 176 μm. Further GradCAM analysis revealed that the predictions were primarily driven by subcortical vascular structures. Finally, we assessed the robustness of our method in a cortical stroke where the brain vasculature is severely impaired. Remarkably, no specific increase in the number of misclassifications was observed, confirming the method's reliability in challenging conditions. Overall, our framework provides accurate and flexible positioning, not relying on a pre-registered reference but rather on conserved vascular patterns.
Collapse
Affiliation(s)
- Théo Lambert
- Neuro-Electronics Research Flanders, Leuven, Belgium
- VIB, Leuven, Belgium
- Imec, Leuven, Belgium
- Department of Neuroscience, Faculty of Medicine, KU Leuven, Leuven, Belgium
| | - Clément Brunner
- Neuro-Electronics Research Flanders, Leuven, Belgium
- VIB, Leuven, Belgium
- Imec, Leuven, Belgium
- Department of Neuroscience, Faculty of Medicine, KU Leuven, Leuven, Belgium
| | - Dries Kil
- Neuro-Electronics Research Flanders, Leuven, Belgium
- VIB, Leuven, Belgium
- Imec, Leuven, Belgium
- Department of Neuroscience, Faculty of Medicine, KU Leuven, Leuven, Belgium
| | | | | | - Gabriel Montaldo
- Neuro-Electronics Research Flanders, Leuven, Belgium
- VIB, Leuven, Belgium
- Imec, Leuven, Belgium
- Department of Neuroscience, Faculty of Medicine, KU Leuven, Leuven, Belgium
| | - Alan Urban
- Neuro-Electronics Research Flanders, Leuven, Belgium
- VIB, Leuven, Belgium
- Imec, Leuven, Belgium
- Department of Neuroscience, Faculty of Medicine, KU Leuven, Leuven, Belgium
| |
Collapse
|
10
|
Azad R, Kazerouni A, Heidari M, Aghdam EK, Molaei A, Jia Y, Jose A, Roy R, Merhof D. Advances in medical image analysis with vision Transformers: A comprehensive review. Med Image Anal 2024; 91:103000. [PMID: 37883822 DOI: 10.1016/j.media.2023.103000] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2023] [Revised: 09/30/2023] [Accepted: 10/11/2023] [Indexed: 10/28/2023]
Abstract
The remarkable performance of the Transformer architecture in natural language processing has recently also triggered broad interest in Computer Vision. Among other merits, Transformers are witnessed as capable of learning long-range dependencies and spatial correlations, which is a clear advantage over convolutional neural networks (CNNs), which have been the de facto standard in Computer Vision problems so far. Thus, Transformers have become an integral part of modern medical image analysis. In this review, we provide an encyclopedic review of the applications of Transformers in medical imaging. Specifically, we present a systematic and thorough review of relevant recent Transformer literature for different medical image analysis tasks, including classification, segmentation, detection, registration, synthesis, and clinical report generation. For each of these applications, we investigate the novelty, strengths and weaknesses of the different proposed strategies and develop taxonomies highlighting key properties and contributions. Further, if applicable, we outline current benchmarks on different datasets. Finally, we summarize key challenges and discuss different future research directions. In addition, we have provided cited papers with their corresponding implementations in https://github.com/mindflow-institue/Awesome-Transformer.
Collapse
Affiliation(s)
- Reza Azad
- Faculty of Electrical Engineering and Information Technology, RWTH Aachen University, Aachen, Germany
| | - Amirhossein Kazerouni
- School of Electrical Engineering, Iran University of Science and Technology, Tehran, Iran
| | - Moein Heidari
- School of Electrical Engineering, Iran University of Science and Technology, Tehran, Iran
| | | | - Amirali Molaei
- School of Computer Engineering, Iran University of Science and Technology, Tehran, Iran
| | - Yiwei Jia
- Faculty of Electrical Engineering and Information Technology, RWTH Aachen University, Aachen, Germany
| | - Abin Jose
- Faculty of Electrical Engineering and Information Technology, RWTH Aachen University, Aachen, Germany
| | - Rijo Roy
- Faculty of Electrical Engineering and Information Technology, RWTH Aachen University, Aachen, Germany
| | - Dorit Merhof
- Faculty of Informatics and Data Science, University of Regensburg, Regensburg, Germany; Fraunhofer Institute for Digital Medicine MEVIS, Bremen, Germany.
| |
Collapse
|
11
|
Gao X, Zheng G. SMILE: Siamese Multi-scale Interactive-representation LEarning for Hierarchical Diffeomorphic Deformable image registration. Comput Med Imaging Graph 2024; 111:102322. [PMID: 38157671 DOI: 10.1016/j.compmedimag.2023.102322] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2023] [Revised: 11/23/2023] [Accepted: 12/14/2023] [Indexed: 01/03/2024]
Abstract
Deformable medical image registration plays an important role in many clinical applications. It aims to find a dense deformation field to establish point-wise correspondences between a pair of fixed and moving images. Recently, unsupervised deep learning-based registration methods have drawn more and more attention because of fast inference at testing stage. Despite remarkable progress, existing deep learning-based methods suffer from several limitations including: (a) they often overlook the explicit modeling of feature correspondences due to limited receptive fields; (b) the performance on image pairs with large spatial displacements is still limited since the dense deformation field is regressed from features learned by local convolutions; and (c) desirable properties, including topology-preservation and the invertibility of transformation, are often ignored. To address above limitations, we propose a novel Convolutional Neural Network (CNN) consisting of a Siamese Multi-scale Interactive-representation LEarning (SMILE) encoder and a Hierarchical Diffeomorphic Deformation (HDD) decoder. Specifically, the SMILE encoder aims for effective feature representation learning and spatial correspondence establishing while the HDD decoder seeks to regress the dense deformation field in a coarse-to-fine manner. We additionally propose a novel Local Invertible Loss (LIL) to encourage topology-preservation and local invertibility of the regressed transformation while keeping high registration accuracy. Extensive experiments conducted on two publicly available brain image datasets demonstrate the superiority of our method over the state-of-the-art (SOTA) approaches. Specifically, on the Neurite-OASIS dataset, our method achieved an average DSC of 0.815 and an average ASSD of 0.633 mm.
Collapse
Affiliation(s)
- Xiaoru Gao
- Institute of Medical Robotics, School of Biomedical Engineering, 800 DongChuan Road, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Guoyan Zheng
- Institute of Medical Robotics, School of Biomedical Engineering, 800 DongChuan Road, Shanghai Jiao Tong University, Shanghai, 200240, China.
| |
Collapse
|
12
|
Hiep MAJ, Heerink WJ, Groen HC, Ruers TJM. Feasibility of tracked ultrasound registration for pelvic-abdominal tumor navigation: a patient study. Int J Comput Assist Radiol Surg 2023; 18:1725-1734. [PMID: 37227572 DOI: 10.1007/s11548-023-02937-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2023] [Accepted: 04/24/2023] [Indexed: 05/26/2023]
Abstract
PURPOSE Surgical navigation techniques can guide surgeons in localizing pelvic-abdominal malignancies. For abdominal navigation, accurate patient registration is crucial and is generally performed using an intra-operative cone-beam CT (CBCT). However, this method causes 15-min surgical preparation workflow interruption and radiation exposure, and more importantly, it cannot be repeated during surgery to compensate for large patient movement. As an alternative, the accuracy and feasibility of tracked ultrasound (US) registration are assessed in this patient study. METHODS Patients scheduled for surgical navigation during laparotomy of pelvic-abdominal malignancies were prospectively included. In the operating room, two percutaneous tracked US scans of the pelvic bone were acquired: one in supine and one in Trendelenburg patient position. Postoperatively, the bone surface was semiautomatically segmented from US images and registered to the bone surface on the preoperative CT scan. The US registration accuracy was computed using the CBCT registration as a reference and acquisition times were compared. Additionally, both US measurements were compared to quantify the registration error caused by patient movement into Trendelenburg. RESULTS In total, 18 patients were included and analyzed. US registration resulted in a mean surface registration error of 1.2 ± 0.2 mm and a mean target registration error of 3.3 ± 1.4 mm. US acquisitions were 4 × faster than the CBCT scans (two-sample t-test P < 0.05) and could even be performed during standard patient preparation before skin incision. Patient repositioning in Trendelenburg caused a mean target registration error of 7.7 ± 3.3 mm, mainly in cranial direction. CONCLUSION US registration based on the pelvic bone is accurate, fast and feasible for surgical navigation. Further optimization of the bone segmentation algorithm will allow for real-time registration in the clinical workflow. In the end, this would allow intra-operative US registration to correct for large patient movement. TRIAL REGISTRATION This study is registered in ClinicalTrials.gov (NCT05637359).
Collapse
Affiliation(s)
- M A J Hiep
- Department of Surgical Oncology, Netherlands Cancer Institute, 1066 CX, Amsterdam, The Netherlands.
| | - W J Heerink
- Department of Surgical Oncology, Netherlands Cancer Institute, 1066 CX, Amsterdam, The Netherlands
| | - H C Groen
- Department of Surgical Oncology, Netherlands Cancer Institute, 1066 CX, Amsterdam, The Netherlands
| | - T J M Ruers
- Department of Surgical Oncology, Netherlands Cancer Institute, 1066 CX, Amsterdam, The Netherlands
- Faculty of Science and Technology (TNW), Nanobiophysics Group (NBP), University of Twente, 7500 AE, Enschede, The Netherlands
| |
Collapse
|
13
|
Chi Y, Xu Y, Liu H, Wu X, Liu Z, Mao J, Xu G, Huang W. A two-step deep learning method for 3DCT-2DUS kidney registration during breathing. Sci Rep 2023; 13:12846. [PMID: 37553480 PMCID: PMC10409729 DOI: 10.1038/s41598-023-40133-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2023] [Accepted: 08/04/2023] [Indexed: 08/10/2023] Open
Abstract
This work proposed KidneyRegNet, a novel deep registration pipeline for 3D CT and 2D U/S kidney scans of free breathing, which comprises a feature network, and a 3D-2D CNN-based registration network. The feature network has handcrafted texture feature layers to reduce the semantic gap. The registration network is an encoder-decoder structure with loss of feature-image-motion (FIM), which enables hierarchical regression at decoder layers and avoids multiple network concatenation. It was first pretrained with a retrospective dataset cum training data generation strategy and then adapted to specific patient data under unsupervised one-cycle transfer learning in onsite applications. The experiment was performed on 132 U/S sequences, 39 multiple-phase CT and 210 public single-phase CT images, and 25 pairs of CT and U/S sequences. This resulted in a mean contour distance (MCD) of 0.94 mm between kidneys on CT and U/S images and MCD of 1.15 mm on CT and reference CT images. Datasets with small transformations resulted in MCDs of 0.82 and 1.02 mm, respectively. Large transformations resulted in MCDs of 1.10 and 1.28 mm, respectively. This work addressed difficulties in 3DCT-2DUS kidney registration during free breathing via novel network structures and training strategies.
Collapse
Affiliation(s)
- Yanling Chi
- Institute for Infocomm Research (I2R), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way #21-01 Connexis South, Singapore, 138632, Republic of Singapore.
| | - Yuyu Xu
- Key Laboratory of Biological Targeting Diagnosis, Therapy and Rehabilitation of Guangdong Higher Education Institutes, The Fifth Affiliated Hospital of Guangzhou Medical University, Guangzhou, 510700, People's Republic of China
| | - Huiying Liu
- Institute for Infocomm Research (I2R), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way #21-01 Connexis South, Singapore, 138632, Republic of Singapore
| | - Xiaoxiang Wu
- Key Laboratory of Biological Targeting Diagnosis, Therapy and Rehabilitation of Guangdong Higher Education Institutes, The Fifth Affiliated Hospital of Guangzhou Medical University, Guangzhou, 510700, People's Republic of China
| | - Zhiqiang Liu
- Key Laboratory of Biological Targeting Diagnosis, Therapy and Rehabilitation of Guangdong Higher Education Institutes, The Fifth Affiliated Hospital of Guangzhou Medical University, Guangzhou, 510700, People's Republic of China
| | - Jiawei Mao
- Creative Medtech Solutions Pte Ltd, Singapore, Republic of Singapore
| | - Guibin Xu
- Key Laboratory of Biological Targeting Diagnosis, Therapy and Rehabilitation of Guangdong Higher Education Institutes, The Fifth Affiliated Hospital of Guangzhou Medical University, Guangzhou, 510700, People's Republic of China.
| | - Weimin Huang
- Institute for Infocomm Research (I2R), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way #21-01 Connexis South, Singapore, 138632, Republic of Singapore.
| |
Collapse
|
14
|
Shamshad F, Khan S, Zamir SW, Khan MH, Hayat M, Khan FS, Fu H. Transformers in medical imaging: A survey. Med Image Anal 2023; 88:102802. [PMID: 37315483 DOI: 10.1016/j.media.2023.102802] [Citation(s) in RCA: 88] [Impact Index Per Article: 88.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2022] [Revised: 03/11/2023] [Accepted: 03/23/2023] [Indexed: 06/16/2023]
Abstract
Following unprecedented success on the natural language tasks, Transformers have been successfully applied to several computer vision problems, achieving state-of-the-art results and prompting researchers to reconsider the supremacy of convolutional neural networks (CNNs) as de facto operators. Capitalizing on these advances in computer vision, the medical imaging field has also witnessed growing interest for Transformers that can capture global context compared to CNNs with local receptive fields. Inspired from this transition, in this survey, we attempt to provide a comprehensive review of the applications of Transformers in medical imaging covering various aspects, ranging from recently proposed architectural designs to unsolved issues. Specifically, we survey the use of Transformers in medical image segmentation, detection, classification, restoration, synthesis, registration, clinical report generation, and other tasks. In particular, for each of these applications, we develop taxonomy, identify application-specific challenges as well as provide insights to solve them, and highlight recent trends. Further, we provide a critical discussion of the field's current state as a whole, including the identification of key challenges, open problems, and outlining promising future directions. We hope this survey will ignite further interest in the community and provide researchers with an up-to-date reference regarding applications of Transformer models in medical imaging. Finally, to cope with the rapid development in this field, we intend to regularly update the relevant latest papers and their open-source implementations at https://github.com/fahadshamshad/awesome-transformers-in-medical-imaging.
Collapse
Affiliation(s)
- Fahad Shamshad
- MBZ University of Artificial Intelligence, Abu Dhabi, United Arab Emirates.
| | - Salman Khan
- MBZ University of Artificial Intelligence, Abu Dhabi, United Arab Emirates; CECS, Australian National University, Canberra ACT 0200, Australia
| | - Syed Waqas Zamir
- Inception Institute of Artificial Intelligence, Abu Dhabi, United Arab Emirates
| | | | - Munawar Hayat
- Faculty of IT, Monash University, Clayton VIC 3800, Australia
| | - Fahad Shahbaz Khan
- MBZ University of Artificial Intelligence, Abu Dhabi, United Arab Emirates; Computer Vision Laboratory, Linköping University, Sweden
| | - Huazhu Fu
- Institute of High Performance Computing, Agency for Science, Technology and Research (A*STAR), Singapore
| |
Collapse
|
15
|
Kim M, Chung M, Shin YG, Kim B. Automatic registration of dental CT and 3D scanned model using deep split jaw and surface curvature. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 233:107467. [PMID: 36921464 DOI: 10.1016/j.cmpb.2023.107467] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/13/2022] [Revised: 02/07/2023] [Accepted: 03/04/2023] [Indexed: 06/18/2023]
Abstract
BACKGROUND AND OBJECTIVES In the medical field, various image registration applications have been studied. In dentistry, the registration of computed tomography (CT) volume data and 3D optically scanned models is essential for various clinical applications, including orthognathic surgery, implant surgical planning, and augmented reality. Our purpose was to present a fully automatic registration method of dental CT data and 3D scanned models. METHODS We use a 2D convolutional neural network to regress a curve splitting the maxilla (i.e., upper jaw) and mandible (i.e., lower jaw) and the points specifying the front and back ends of the crown from the CT data. Using this regressed information, we extract the point cloud and vertices corresponding to the tooth crown from the CT and scanned data, respectively. We introduce a novel metric, called curvature variance of neighbor (CVN), to discriminate between highly fluctuating and smoothly varying regions of the tooth crown. The registration based on CVN enables more accurate fine registration while reducing the effects of metal artifacts. Moreover, the proposed method does not require any preprocessing such as extracting the iso-surface for the tooth crown from the CT data, thereby significantly reducing the computation time. RESULTS We evaluated the proposed method with the comparison to several promising registration techniques. Our experimental results using three datasets demonstrated that the proposed method exhibited higher registration accuracy (i.e., 2.85, 1.92, and 7.73 times smaller distance errors for individual datasets) and smaller computation time (i.e., 4.12 times faster registration) than one of the state-of-the-art methods. Moreover, the proposed method worked considerably well for partially scanned data, whereas other methods suffered from the unbalancing of information between the CT and scanned data. CONCLUSIONS The proposed method was able to perform fully automatic and highly accurate registration of dental CT data and 3D scanned models, even with severe metal artifacts. In addition, it could achieve fast registration because it did not require any preprocessing for iso-surface reconstruction from the CT data.
Collapse
Affiliation(s)
- Minchang Kim
- Department of Computer Science and Engineering, Seoul National University, 1 Gwanak-ro, Gwanak-gu, Seoul 08826, Republic of Korea
| | - Minyoung Chung
- School of Software, Soongsil University, 369 Sangdo-Ro, Dongjak-Gu, Seoul 06978, Republic of Korea
| | - Yeong-Gil Shin
- Department of Computer Science and Engineering, Seoul National University, 1 Gwanak-ro, Gwanak-gu, Seoul 08826, Republic of Korea
| | - Bohyoung Kim
- Division of Biomedical Engineering, Hankuk University of Foreign Studies, 81 Oedae-ro, Mohyeon-myeon, Cheoin-gu, Yongin-si, Gyeonggi-do 17035, Republic of Korea.
| |
Collapse
|
16
|
Pérez de Frutos J, Pedersen A, Pelanis E, Bouget D, Survarachakan S, Langø T, Elle OJ, Lindseth F. Learning deep abdominal CT registration through adaptive loss weighting and synthetic data generation. PLoS One 2023; 18:e0282110. [PMID: 36827289 PMCID: PMC9956065 DOI: 10.1371/journal.pone.0282110] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2022] [Accepted: 02/08/2023] [Indexed: 02/25/2023] Open
Abstract
PURPOSE This study aims to explore training strategies to improve convolutional neural network-based image-to-image deformable registration for abdominal imaging. METHODS Different training strategies, loss functions, and transfer learning schemes were considered. Furthermore, an augmentation layer which generates artificial training image pairs on-the-fly was proposed, in addition to a loss layer that enables dynamic loss weighting. RESULTS Guiding registration using segmentations in the training step proved beneficial for deep-learning-based image registration. Finetuning the pretrained model from the brain MRI dataset to the abdominal CT dataset further improved performance on the latter application, removing the need for a large dataset to yield satisfactory performance. Dynamic loss weighting also marginally improved performance, all without impacting inference runtime. CONCLUSION Using simple concepts, we improved the performance of a commonly used deep image registration architecture, VoxelMorph. In future work, our framework, DDMR, should be validated on different datasets to further assess its value.
Collapse
Affiliation(s)
| | - André Pedersen
- Department of Health Research, SINTEF, Trondheim, Norway
- Department of Clinical and Molecular Medicine, Norwegian University of Technology (NTNU), Trondheim, Norway
- Clinic of Surgery, St. Olavs hospital, Trondheim University Hospital, Trondheim, Norway
| | | | - David Bouget
- Department of Health Research, SINTEF, Trondheim, Norway
| | | | - Thomas Langø
- Department of Health Research, SINTEF, Trondheim, Norway
- Research Department, Future Operating Room, St. Olavs hospital, Trondheim University Hospital, Trondheim, Norway
| | - Ole-Jakob Elle
- The Intervention Centre, Oslo University Hospital, Oslo, Norway
| | - Frank Lindseth
- Department of Computer Science, Norwegian University of Science and Technology (NTNU), Trondheim, Norway
| |
Collapse
|
17
|
Role of Ensemble Deep Learning for Brain Tumor Classification in Multiple Magnetic Resonance Imaging Sequence Data. Diagnostics (Basel) 2023; 13:diagnostics13030481. [PMID: 36766587 PMCID: PMC9914433 DOI: 10.3390/diagnostics13030481] [Citation(s) in RCA: 13] [Impact Index Per Article: 13.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/29/2022] [Revised: 01/24/2023] [Accepted: 01/26/2023] [Indexed: 01/31/2023] Open
Abstract
The biopsy is a gold standard method for tumor grading. However, due to its invasive nature, it has sometimes proved fatal for brain tumor patients. As a result, a non-invasive computer-aided diagnosis (CAD) tool is required. Recently, many magnetic resonance imaging (MRI)-based CAD tools have been proposed for brain tumor grading. The MRI has several sequences, which can express tumor structure in different ways. However, a suitable MRI sequence for brain tumor classification is not yet known. The most common brain tumor is 'glioma', which is the most fatal form. Therefore, in the proposed study, to maximize the classification ability between low-grade versus high-grade glioma, three datasets were designed comprising three MRI sequences: T1-Weighted (T1W), T2-weighted (T2W), and fluid-attenuated inversion recovery (FLAIR). Further, five well-established convolutional neural networks, AlexNet, VGG16, ResNet18, GoogleNet, and ResNet50 were adopted for tumor classification. An ensemble algorithm was proposed using the majority vote of above five deep learning (DL) models to produce more consistent and improved results than any individual model. Five-fold cross validation (K5-CV) protocol was adopted for training and testing. For the proposed ensembled classifier with K5-CV, the highest test accuracies of 98.88 ± 0.63%, 97.98 ± 0.86%, and 94.75 ± 0.61% were achieved for FLAIR, T2W, and T1W-MRI data, respectively. FLAIR-MRI data was found to be most significant for brain tumor classification, where it showed a 4.17% and 0.91% improvement in accuracy against the T1W-MRI and T2W-MRI sequence data, respectively. The proposed ensembled algorithm (MajVot) showed significant improvements in the average accuracy of three datasets of 3.60%, 2.84%, 1.64%, 4.27%, and 1.14%, respectively, against AlexNet, VGG16, ResNet18, GoogleNet, and ResNet50.
Collapse
|
18
|
Bhimanpallewar RN, Khan SI, Raj KB, Gulati K, Bhasin N, Raj R. Federate learning on Web browsing data with statically and machine learning technique. INTERNATIONAL JOURNAL OF PERVASIVE COMPUTING AND COMMUNICATIONS 2022. [DOI: 10.1108/ijpcc-05-2022-0184] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Purpose
Federation analytics approaches are a present area of study that has already progressed beyond the analysis of metrics and counts. It is possible to acquire aggregated information about on-device data by training machine learning models using federated learning techniques without any of the raw data ever having to leave the devices in the issue. Web browser forensics research has been focused on individual Web browsers or architectural analysis of specific log files rather than on broad topics. This paper aims to propose major tools used for Web browser analysis.
Design/methodology/approach
Each kind of Web browser has its own unique set of features. This allows the user to choose their preferred browsers or to check out many browsers at once. If a forensic examiner has access to just one Web browser's log files, he/she makes it difficult to determine which sites a person has visited. The agent must thus be capable of analyzing all currently available Web browsers on a single workstation and doing an integrated study of various Web browsers.
Findings
Federated learning has emerged as a training paradigm in such settings. Web browser forensics research in general has focused on certain browsers or the computational modeling of specific log files. Internet users engage in a wide range of activities using an internet browser, such as searching for information and sending e-mails.
Originality/value
It is also essential that the investigator have access to user activity when conducting an inquiry. This data, which may be used to assess information retrieval activities, is very critical. In this paper, the authors purposed a major tool used for Web browser analysis. This study's proposed algorithm is capable of protecting data privacy effectively in real-world experiments.
Collapse
|
19
|
He Y, Wang A, Li S, Yang Y, Hao A. Nonfinite-modality data augmentation for brain image registration. Comput Biol Med 2022; 147:105780. [PMID: 35772329 DOI: 10.1016/j.compbiomed.2022.105780] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/18/2022] [Revised: 05/24/2022] [Accepted: 06/19/2022] [Indexed: 01/25/2023]
Abstract
Brain image registration is fundamental for brain medical image analysis. However, the lack of paired images with diverse modalities and corresponding ground truth deformations for training hinder its development. We propose a novel nonfinite-modality data augmentation for brain image registration to combat this. Specifically, some available whole-brain segmentation masks, including complete fine brain anatomical structures, are collected from the actual brain dataset, OASIS-3. One whole-brain segmentation mask can generate many nonfinite-modality brain images by randomly merging some fine anatomical structures and subsequently sampling the intensities for each fine anatomical structure using random Gaussian distribution. Furthermore, to get more realistic deformations as the ground truth, an improved 3D Variational Auto-encoder (VAE) is proposed by introducing the intensity-level reconstruction loss and the structure-level reconstruction loss. Based on the generated images and trained improved 3D VAE, a new Synthetic Nonfinite-Modality Brain Image Dataset (SNMBID) is created. Experiments show that pre-training on SNMBID can improve the accuracy of registration. Notably, SNMBID can be a landmark for evaluating other brain registration methods, and the model trained on the SNMBID can be a baseline for the brain image registration task. Our code is available at https://github.com/MangoWAY/SMIBID_BrainRegistration.
Collapse
Affiliation(s)
- Yuanbo He
- State Key Laboratory of Virtual Reality Technology and Systems, Beihang University, Beijing, 100191, China; Peng Cheng Laboratory, Shenzhen, 518055, China.
| | - Aoyu Wang
- ByteDance Intelligent Creation, Beijing, 100191, China.
| | - Shuai Li
- State Key Laboratory of Virtual Reality Technology and Systems, Beihang University, Beijing, 100191, China; Peng Cheng Laboratory, Shenzhen, 518055, China.
| | - Yikang Yang
- State Key Laboratory of Virtual Reality Technology and Systems, Beihang University, Beijing, 100191, China.
| | - Aimin Hao
- State Key Laboratory of Virtual Reality Technology and Systems, Beihang University, Beijing, 100191, China; Peng Cheng Laboratory, Shenzhen, 518055, China.
| |
Collapse
|
20
|
Federate learning of corporate social authority and industry 4.0 that focus on young people: a strategic management framework for human resources. INTERNATIONAL JOURNAL OF PERVASIVE COMPUTING AND COMMUNICATIONS 2022. [DOI: 10.1108/ijpcc-02-2022-0056] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Purpose
The young population of the globe is defined by individuals aged 15 to 24 years. Based on statistics from the Instituto Brasileiro de Geografia e Estatística (IBGE), the second largest women population among 15 years as well as 19 years was in 2017 only behind 35 and 39 years. At this time, the Brazilian male population was higher. The difficulties of the young generation affected the preceding generation and promoted social dynamism. The worldwide data shows that the generation of young and the digital world have been constantly sought, but in reality, approximately one-third of the population in 2017 had no access to the internet.
Design/methodology/approach
The worldwide movement around topics such as strategy on its threefold basis and Industry 4.0 enable a link to company duty towards society to be established. This present study was produced from 1 March 2020 to 2 September 2020 via resources of human and literature evaluation relating to the idea of strategic, Industry 4.0, the responsibility of society and the creation of youth. Its motive is the global creation of youth. Two recommendations should be made after studying the literature and information gathering that enabled “analyzing social responsibility of the company and industry 4.0 with a pivot on young creation: a strategic framework for resources of human management”.
Findings
The adoption of defensible practices and technology bring forth by the revolution in industrial is emphasized worldwide.
Originality/value
The focus on the usage of these ideas is essential, so that young people can absorb the workforce in the labour market. To achieve this, the CSR idea combines this theoretical triple-created recent study.
Collapse
|
21
|
Ma L, Liang H, Han B, Yang S, Zhang X, Liao H. Augmented reality navigation with ultrasound-assisted point cloud registration for percutaneous ablation of liver tumors. Int J Comput Assist Radiol Surg 2022; 17:1543-1552. [PMID: 35704238 DOI: 10.1007/s11548-022-02671-7] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2022] [Accepted: 05/02/2022] [Indexed: 11/28/2022]
Abstract
PURPOSE We present a novel augmented reality (AR) surgical navigation method with ultrasound-assisted point cloud registration for percutaneous ablation of liver tumors. A preliminary study is carried out to verify its feasibility. METHODS Two three-dimensional (3D) point clouds of the liver surface are derived from the preoperative images and intraoperative tracked US images, respectively. To compensate for the soft tissue deformation, the point cloud registration between the preoperative images and the liver is performed using the non-rigid iterative closest point (ICP) algorithm. A 3D AR device based on integral videography technology is designed to accurately display naked-eye 3D images for surgical navigation. Based on the above registration, naked-eye 3D images of the liver surface, planning path, entry points, and tumor can be overlaid in situ through our 3D AR device. Finally, the AR-guided targeting accuracy is evaluated through entry point positioning. RESULTS Experiments on both the liver phantom and in vitro pork liver were conducted. Several entry points on the liver surface were used to evaluate the targeting accuracy. The preliminary validation on the liver phantom showed average entry-point errors (EPEs) of 2.34 ± 0.45 mm, 2.25 ± 0.72 mm, 2.71 ± 0.82 mm, and 2.50 ± 1.11 mm at distinct US point cloud coverage rates of 100%, 75%, 50%, and 25%, respectively. The average EPEs of the deformed pork liver were 4.49 ± 1.88 mm and 5.02 ± 2.03 mm at the coverage rates of 100% and 75%, and the average covered-entry-point errors (CEPEs) were 4.96 ± 2.05 mm and 2.97 ± 1.37 mm at 50% and 25%, respectively. CONCLUSION Experimental outcomes demonstrate that the proposed AR navigation method based on US-assisted point cloud registration has achieved an acceptable targeting accuracy on the liver surface even in the case of liver deformation.
Collapse
Affiliation(s)
- Longfei Ma
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, 100084, China
| | - Hanying Liang
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, 100084, China
| | - Boxuan Han
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, 100084, China
| | - Shizhong Yang
- Hepato-Pancreato-Biliary Center, Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, 102218, China
| | - Xinran Zhang
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, 100084, China
| | - Hongen Liao
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, 100084, China.
| |
Collapse
|
22
|
Nguyen HP, Kim T, Kim S. Markerless registration approach using dynamic touchable region model. Int J Med Robot 2022; 18:e2376. [DOI: 10.1002/rcs.2376] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2021] [Revised: 01/27/2022] [Accepted: 01/29/2022] [Indexed: 11/09/2022]
Affiliation(s)
- Hang Phuong Nguyen
- Department of Electrical, Electronic, and Computer Engineering University of Ulsan Ulsan South Korea
| | - Taeho Kim
- Department of Electrical, Electronic, and Computer Engineering University of Ulsan Ulsan South Korea
| | - Sungmin Kim
- Department of Electrical, Electronic, and Computer Engineering University of Ulsan Ulsan South Korea
| |
Collapse
|
23
|
Reattachable fiducial skin marker for automatic multimodality registration. Int J Comput Assist Radiol Surg 2022; 17:2141-2150. [PMID: 35604488 PMCID: PMC9515062 DOI: 10.1007/s11548-022-02639-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2022] [Accepted: 04/08/2022] [Indexed: 11/05/2022]
Abstract
Abstract
Purpose
Fusing image information has become increasingly important for optimal diagnosis and treatment of the patient. Despite intensive research towards markerless registration approaches, fiducial marker-based methods remain the default choice for a wide range of applications in clinical practice. However, as especially non-invasive markers cannot be positioned reproducibly in the same pose on the patient, pre-interventional imaging has to be performed immediately before the intervention for fiducial marker-based registrations.
Methods
We propose a new non-invasive, reattachable fiducial skin marker concept for multi-modal registration approaches including the use of electromagnetic or optical tracking technologies. We furthermore describe a robust, automatic fiducial marker localization algorithm for computed tomography (CT) and magnetic resonance imaging (MRI) images. Localization of the new fiducial marker has been assessed for different marker configurations using both CT and MRI. Furthermore, we applied the marker in an abdominal phantom study. For this, we attached the marker at three poses to the phantom, registered ten segmented targets of the phantom’s CT image to live ultrasound images and determined the target registration error (TRE) for each target and each marker pose.
Results
Reattachment of the marker was possible with a mean precision of 0.02 mm ± 0.01 mm. Our algorithm successfully localized the marker automatically in all ($$n=201$$
n
=
201
) evaluated CT/MRI images. Depending on the marker pose, the mean ($$n=10$$
n
=
10
) TRE of the abdominal phantom study ranged from 1.51 ± 0.75 mm to 4.65 ± 1.22 mm.
Conclusions
The non-invasive, reattachable skin marker concept allows reproducible positioning of the marker and automatic localization in different imaging modalities. The low TREs indicate the potential applicability of the marker concept for clinical interventions, such as the puncture of abdominal lesions, where current image-based registration approaches still lack robustness and existing marker-based methods are often impractical.
Collapse
|
24
|
2D/3D Multimode Medical Image Registration Based on Normalized Cross-Correlation. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12062828] [Citation(s) in RCA: 40] [Impact Index Per Article: 20.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/15/2022]
Abstract
Image-guided surgery (IGS) can reduce the risk of tissue damage and improve the accuracy and targeting of lesions by increasing the surgery’s visual field. Three-dimensional (3D) medical images can provide spatial location information to determine the location of lesions and plan the operation process. For real-time tracking and adjusting the spatial position of surgical instruments, two-dimensional (2D) images provide real-time intraoperative information. In this experiment, 2D/3D medical image registration algorithm based on the gray level is studied, and the registration based on normalized cross-correlation is realized. The Gaussian Laplacian second-order differential operator is introduced as a new similarity measure to increase edge information and internal detail information to solve single information and small convergence regions of the normalized cross-correlation algorithm. The multiresolution strategy improves the registration accuracy and efficiency to solve the low efficiency of the normalized cross-correlation algorithm.
Collapse
|
25
|
Ringel MJ, Richey WL, Heiselman J, Luo M, Meszoely IM, Miga MI. Breast image registration for surgery: Insights on material mechanics modeling. PROCEEDINGS OF SPIE--THE INTERNATIONAL SOCIETY FOR OPTICAL ENGINEERING 2022; 12034:1203411. [PMID: 35607388 PMCID: PMC9124453 DOI: 10.1117/12.2611787] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Breast conserving surgery (BCS) is a common procedure for early-stage breast cancer patients. Supine preoperative magnetic resonance (MR) breast imaging for visualizing tumor location and extent, while not standard for procedural guidance, more closely represents the surgical presentation compared to conventional diagnostic pendant positioning. Optimal utilization for surgical guidance, however, requires a fast and accurate image-to-physical registration from preoperative imaging to intraoperative surgical presentation. In this study, three registration methods were investigated on healthy volunteers' breasts (n=11) with the arm-down position simulating preoperative imaging and arm-up position simulating intraoperative data. The registration methods included: (1) point-based rigid registration using synthetic fiducials, (2) non-rigid biomechanical model-based registration using sparse data, and (3) a data-dense 3D diffeomorphic image-based registration from the Advanced Normalization Tools (ANTs) repository. The average target registration errors (TRE) were 10.4 ± 2.3, 6.4 ± 1.5, and 2.8 ± 1.3 mm (mean ± standard deviation) and the average fiducial registration errors (FRE) were 7.8 ± 1.7, 2.5 ± 1.1, and 3.1 ± 1.1 mm (mean ± standard deviation) for the point-based rigid, nonrigid biomechanical, and ANTs registrations, respectively. Additionally, common mechanics-based deformation metrics (volume change and anisotropy) were calculated from the ANTs deformation field. The average metrics revealed anisotropic tissue behavior and a statistical difference in volume change between glandular and adipose tissue, suggesting that nonrigid modeling methods may be improved by incorporating material heterogeneity and anisotropy. Overall, registration accuracy significantly improved with increasingly flexible registration methods, which may inform future development of image guidance systems for lumpectomy procedures.
Collapse
Affiliation(s)
- Morgan J Ringel
- Vanderbilt University, Department of Biomedical Engineering, Nashville, TN USA
- Vanderbilt Institute for Surgery and Engineering, Nashville, TN USA
| | - Winona L Richey
- Vanderbilt University, Department of Biomedical Engineering, Nashville, TN USA
- Vanderbilt Institute for Surgery and Engineering, Nashville, TN USA
| | - Jon Heiselman
- Vanderbilt University, Department of Biomedical Engineering, Nashville, TN USA
- Vanderbilt Institute for Surgery and Engineering, Nashville, TN USA
| | - Ma Luo
- Vanderbilt University, Department of Biomedical Engineering, Nashville, TN USA
- Vanderbilt Institute for Surgery and Engineering, Nashville, TN USA
| | - Ingrid M Meszoely
- Vanderbilt University Medical Center, Division of Surgical Oncology, Nashville, TN USA
| | - Michael I Miga
- Vanderbilt University, Department of Biomedical Engineering, Nashville, TN USA
- Vanderbilt University Department of Radiology and Radiological Sciences, Nashville, TN USA
- Vanderbilt Institute for Surgery and Engineering, Nashville, TN USA
- Vanderbilt University Medical Center, Department of Neurological Surgery, Nashville, TN USA
- Vanderbilt University Medical Center, Department of Otolaryngology-Head and Neck Surgery, Nashville, TN USA
| |
Collapse
|
26
|
Musa Jaber M, Yussof S, S. Elameer A, Yeng Weng L, Khalil Abd S, Nayyar A. Medical Image Analysis Using Deep Learning and Distribution Pattern Matching Algorithm. COMPUTERS, MATERIALS & CONTINUA 2022; 72:2175-2190. [DOI: 10.32604/cmc.2022.023387] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/06/2021] [Accepted: 12/31/2021] [Indexed: 09/02/2023]
|
27
|
A Medical Image Registration Method Based on Progressive Images. COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE 2021; 2021:4504306. [PMID: 34367316 PMCID: PMC8337131 DOI: 10.1155/2021/4504306] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/16/2021] [Accepted: 07/03/2021] [Indexed: 01/26/2023]
Abstract
Background Medical image registration is an essential task for medical image analysis in various applications. In this work, we develop a coarse-to-fine medical image registration method based on progressive images and SURF algorithm (PI-SURF) for higher registration accuracy. Methods As a first step, the reference image and the floating image are fused to generate multiple progressive images. Thereafter, the floating image and progressive image are registered to get the coarse registration result based on the SURF algorithm. For further improvement, the coarse registration result and the reference image are registered to perform fine image registration. The appropriate progressive image has been investigated by experiments. The mutual information (MI), normal mutual information (NMI), normalized correlation coefficient (NCC), and mean square difference (MSD) similarity metrics are used to demonstrate the potential of the PI-SURF method. Results For the unimodal and multimodal registration, the PI-SURF method achieves the best results compared with the mutual information method, Demons method, Demons+B-spline method, and SURF method. The MI, NMI, and NCC of PI-SURF are improved by 15.5%, 1.31%, and 7.3%, respectively, while MSD decreased by 13.2% for the multimodal registration compared with the optimal result of the state-of-the-art methods. Conclusions The extensive experiments show that the proposed PI-SURF method achieves higher quality of registration.
Collapse
|
28
|
Chen J, Yuan F, Shen Y, Wang J. Multimodality-based knee joint modelling method with bone and cartilage structures for total knee arthroplasty. Int J Med Robot 2021; 17:e2316. [PMID: 34312966 DOI: 10.1002/rcs.2316] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2021] [Revised: 07/19/2021] [Accepted: 07/22/2021] [Indexed: 11/11/2022]
Abstract
OBJECTIVE We propose a robust and accurate knee joint modelling method with bone and cartilage structures to enable accurate surgical guidance for knee surgery. METHODS A multimodality registration strategy is proposed to fuse magnetic resonance (MR) and computed tomography (CT) images of the femur and tibia separately to remove spatial inconsistency caused by knee bending in CT/MR scans. Automatic segmentation of the femur, tibia and cartilages is carried out with region of interest clustering and intensity analysis based on the multimodal fusion of images. RESULTS Experimental results show that the registration error is 1.13 ± 0.30 mm. The Dice similarity coefficient values of the proposed segmentation method of the femur, tibia, femoral and tibial cartilages are 0.969, 0.966, 0.910 and 0.872, respectively. CONCLUSIONS This study demonstrates the feasibility and effectiveness of multimodality-based registration and segmentation methods for knee joint modelling. The proposed method can provide users with 3D anatomical models of the femur, tibia, and cartilages with few human inputs.
Collapse
Affiliation(s)
- Jiahe Chen
- School of Mechanical Engineering and Automation, Beihang University, Beijing, China
| | - Fuzhen Yuan
- Knee Surgery Department of the Institute of Sports Medicine, Peking University Third Hospital, Beijing, China
| | - Yu Shen
- School of Mechanical Engineering and Automation, Beihang University, Beijing, China
| | - Junchen Wang
- School of Mechanical Engineering and Automation, Beihang University, Beijing, China.,Beijing Advanced Innovation Centre for Biomedical Engineering, Beihang University, Beijing, China
| |
Collapse
|
29
|
Shen L, Ji C, Lin J, Yang H. Construction of Vertebral Body Tracking Algorithm Based on Dynamic Imaging Parameter Measurement and Its Application in the Treatment of Lumbar Instability. JOURNAL OF MEDICAL IMAGING AND HEALTH INFORMATICS 2021. [DOI: 10.1166/jmihi.2021.3534] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
Abstract
Static imaging measurements could not truly reflect the dynamic panorama of the lumbar movement process, and the abnormal activities between the lumbar vertebrae and their dynamic balance could not be observed, resulting in difficulties in the mechanism analysis of lumbar instability
and the efficacy evaluation of manipulation therapy. Therefore, this paper constructed a vertebral tracking algorithm based on dynamic imaging parameter measurement through imaging parameter measurement and calculation. According to the imaging data obtained by vertebral body tracking algorithm,
the corresponding statistical methods were used to compare the functional scores before and after manipulation and the changes of imaging data, so as to evaluate the therapeutic effect of manipulation on lumbar instability. Through the clinical observation and imaging analysis of 15 patients
with lumbar instability before and after manipulation treatment, it is verified that the vertebra tracking algorithm is effective in the vertebra tracking and plays a positive role in the treatment of lumbar instability.
Collapse
Affiliation(s)
- Lanjuan Shen
- The First People's Hospital of Hangzhou City, Zhejiang Province, Hangzhou 310006, Zhejiang, China
| | - Cheng Ji
- The First People's Hospital of Hangzhou City, Zhejiang Province, Hangzhou 310006, Zhejiang, China
| | - Jian Lin
- Xiaoshan, Hangzhou City, Zhejiang Province Hospital, Hangzhou 311201, Zhejiang, China
| | - Hongping Yang
- The First People's Hospital of Hangzhou City, Zhejiang Province, Hangzhou 310006, Zhejiang, China
| |
Collapse
|
30
|
Fiedler C, Jacobs PP, Müller M, Kolbig S, Grunert R, Meixensberger J, Winkler D. A Comparative Study of Automatic Localization Algorithms for Spherical Markers within 3D MRI Data. Brain Sci 2021; 11:brainsci11070876. [PMID: 34208999 PMCID: PMC8301951 DOI: 10.3390/brainsci11070876] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2021] [Revised: 06/23/2021] [Accepted: 06/28/2021] [Indexed: 11/16/2022] Open
Abstract
Localization of features and structures in images is an important task in medical image-processing. Characteristic structures and features are used in diagnostics and surgery planning for spatial adjustments of the volumetric data, including image registration or localization of bone-anchors and fiducials. Since this task is highly recurrent, a fast, reliable and automated approach without human interaction and parameter adjustment is of high interest. In this paper we propose and compare four image processing pipelines, including algorithms for automatic detection and localization of spherical features within 3D MRI data. We developed a convolution based method as well as algorithms based on connected-components labeling and analysis and the circular Hough-transform. A blob detection related approach, analyzing the Hessian determinant, was examined. Furthermore, we introduce a novel spherical MRI-marker design. In combination with the proposed algorithms and pipelines, this allows the detection and spatial localization, including the direction, of fiducials and bone-anchors.
Collapse
Affiliation(s)
- Christian Fiedler
- Department of Neurosurgery, University of Leipzig, 04103 Leipzig, SN, Germany; (C.F.); (R.G.); (J.M.); (D.W.)
- Department of Physical Engineering/Computer Sciences, University of Applied Sciences, 08056 Zwickau, SN, Germany;
| | - Paul-Philipp Jacobs
- Department of Neurosurgery, University of Leipzig, 04103 Leipzig, SN, Germany; (C.F.); (R.G.); (J.M.); (D.W.)
- Correspondence:
| | - Marcel Müller
- Fraunhofer Institute for Machine Tools and Forming Technology, 01187 Dresden, SN, Germany;
| | - Silke Kolbig
- Department of Physical Engineering/Computer Sciences, University of Applied Sciences, 08056 Zwickau, SN, Germany;
| | - Ronny Grunert
- Department of Neurosurgery, University of Leipzig, 04103 Leipzig, SN, Germany; (C.F.); (R.G.); (J.M.); (D.W.)
- Fraunhofer Institute for Machine Tools and Forming Technology, 01187 Dresden, SN, Germany;
| | - Jürgen Meixensberger
- Department of Neurosurgery, University of Leipzig, 04103 Leipzig, SN, Germany; (C.F.); (R.G.); (J.M.); (D.W.)
| | - Dirk Winkler
- Department of Neurosurgery, University of Leipzig, 04103 Leipzig, SN, Germany; (C.F.); (R.G.); (J.M.); (D.W.)
| |
Collapse
|
31
|
Tang N, Fan J, Wang P, Shi G. Microscope integrated optical coherence tomography system combined with augmented reality. OPTICS EXPRESS 2021; 29:9407-9418. [PMID: 33820369 DOI: 10.1364/oe.420375] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/22/2021] [Accepted: 03/03/2021] [Indexed: 06/12/2023]
Abstract
One of the disadvantages in microscope-integrated optical coherence tomography (MI-OCT) systems is that medical images acquired via different modalities are usually displayed independently. Hence, surgeons have to match two-dimensional and three-dimensional images of the same operative region subjectively. In this paper, we propose a simple registration method to overcome this problem by using guided laser points. This method combines augmented reality with an existing MI-OCT system. The basis of our idea is to introduce a guiding laser into the system, which allows us to identify fiducials in microscopic images. At first, the applied voltages of the scanning galvanometer mirror are used to calculate the fiducials' coordinates in an OCT model. After gathering data at the corresponding points' coordinates, the homography matrix and camera parameters are used to superimpose a reconstructed model on microscopic images. After performing experiments with artificial and animal eyes, we successfully obtain two-dimensional microscopic images of scanning regions with depth information. Moreover, the registration error is 0.04 mm, which is within the limits of medical and surgical errors. Our proposed method could have many potential applications in ophthalmic procedures.
Collapse
|
32
|
Extraction and Visualization of Ocular Blood Vessels in 3D Medical Images Based on Geometric Transformation Algorithm. JOURNAL OF HEALTHCARE ENGINEERING 2021. [DOI: 10.1155/2021/5573381] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
Data extraction and visualization of 3D medical images of ocular blood vessels are performed by geometric transformation algorithm, which first performs random resonance response in a global sense to achieve detection of high-contrast coarse blood vessels and then redefines the input signal as a local image shielding the global detection result to achieve enhanced detection of low-contrast microfine vessels and complete multilevel random resonance segmentation detection. Finally, a random resonance detection method for fundus vessels based on scale decomposition is proposed, in which the images are scale decomposed, the high-frequency signals containing detailed information are randomly resonantly enhanced to achieve microfine vessel segmentation detection, and the final vessel segmentation detection results are obtained after fusing the low-frequency image signals. The optimal stochastic resonance response of the nonlinear model of neurons in the global sense is obtained to detect the high-grade intensity signal; then, the input signal is defined as a local image with high-contrast blood vessels removed, and the parameters are optimized before the detection of the low-grade intensity signal. Finally, the multilevel random resonance response is fused to obtain the segmentation results of the fundus retinal vessels. The sensitivity of the multilevel segmentation method proposed in this paper is significantly improved compared with the global random resonance results, indicating that the method proposed in this paper has obvious advantages in the segmentation of vessels with low-intensity levels. The image library was tested, and the experimental results showed that the new method has a better segmentation effect on low-contrast microscopic blood vessels. The new method not only makes full use of the noise for weak signal detection and segmentation but also provides a new idea of how to achieve multilevel segmentation and recognition of medical images.
Collapse
|
33
|
Nie Z, Li C, Liu H, Yang X. Deformable Image Registration Based on Functions of Bounded Generalized Deformation. Int J Comput Vis 2021. [DOI: 10.1007/s11263-021-01439-x] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
34
|
Venet L, Pati S, Feldman MD, Nasrallah MP, Yushkevich P, Bakas S. Accurate and Robust Alignment of Differently Stained Histologic Images Based on Greedy Diffeomorphic Registration. APPLIED SCIENCES (BASEL, SWITZERLAND) 2021; 11:1892. [PMID: 34290888 PMCID: PMC8291745 DOI: 10.3390/app11041892] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 11/21/2022]
Abstract
Histopathologic assessment routinely provides rich microscopic information about tissue structure and disease process. However, the sections used are very thin, and essentially capture only 2D representations of a certain tissue sample. Accurate and robust alignment of sequentially cut 2D slices should contribute to more comprehensive assessment accounting for surrounding 3D information. Towards this end, we here propose a two-step diffeomorphic registration approach that aligns differently stained histology slides to each other, starting with an initial affine step followed by estimating a deformation field. It was quantitatively evaluated on ample (n = 481) and diverse data from the automatic non-rigid histological image registration challenge, where it was awarded the second rank. The obtained results demonstrate the ability of the proposed approach to robustly (average robustness = 0.9898) and accurately (average relative target registration error = 0.2%) align differently stained histology slices of various anatomical sites while maintaining reasonable computational efficiency (<1 min per registration). The method was developed by adapting a general-purpose registration algorithm designed for 3D radiographic scans and achieved consistently accurate results for aligning high-resolution 2D histologic images. Accurate alignment of histologic images can contribute to a better understanding of the spatial arrangement and growth patterns of cells, vessels, matrix, nerves, and immune cell interactions.
Collapse
Affiliation(s)
- Ludovic Venet
- Center for Biomedical Image Computing & Analytics, University of Pennsylvania, Philadelphia, PA 19104, USA
| | - Sarthak Pati
- Center for Biomedical Image Computing & Analytics, University of Pennsylvania, Philadelphia, PA 19104, USA
- Department of Radiology, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA 19104, USA
- Department of Pathology & Laboratory Medicine, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA 19104, USA
| | - Michael D. Feldman
- Department of Pathology & Laboratory Medicine, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA 19104, USA
| | - MacLean P. Nasrallah
- Department of Pathology & Laboratory Medicine, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA 19104, USA
| | - Paul Yushkevich
- Center for Biomedical Image Computing & Analytics, University of Pennsylvania, Philadelphia, PA 19104, USA
- Department of Radiology, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA 19104, USA
| | - Spyridon Bakas
- Center for Biomedical Image Computing & Analytics, University of Pennsylvania, Philadelphia, PA 19104, USA
- Department of Radiology, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA 19104, USA
- Department of Pathology & Laboratory Medicine, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA 19104, USA
| |
Collapse
|
35
|
Chung M, Lee J, Song W, Song Y, Yang IH, Lee J, Shin YG. Automatic Registration Between Dental Cone-Beam CT and Scanned Surface via Deep Pose Regression Neural Networks and Clustered Similarities. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:3900-3909. [PMID: 32746134 DOI: 10.1109/tmi.2020.3007520] [Citation(s) in RCA: 17] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Computerized registration between maxillofacial cone-beam computed tomography (CT) images and a scanned dental model is an essential prerequisite for surgical planning for dental implants or orthognathic surgery. We propose a novel method that performs fully automatic registration between a cone-beam CT image and an optically scanned model. To build a robust and automatic initial registration method, deep pose regression neural networks are applied in a reduced domain (i.e., two-dimensional image). Subsequently, fine registration is performed using optimal clusters. A majority voting system achieves globally optimal transformations while each cluster attempts to optimize local transformation parameters. The coherency of clusters determines their candidacy for the optimal cluster set. The outlying regions in the iso-surface are effectively removed based on the consensus among the optimal clusters. The accuracy of registration is evaluated based on the Euclidean distance of 10 landmarks on a scanned model, which have been annotated by experts in the field. The experiments show that the registration accuracy of the proposed method, measured based on the landmark distance, outperforms the best performing existing method by 33.09%. In addition to achieving high accuracy, our proposed method neither requires human interactions nor priors (e.g., iso-surface extraction). The primary significance of our study is twofold: 1) the employment of lightweight neural networks, which indicates the applicability of neural networks in extracting pose cues that can be easily obtained and 2) the introduction of an optimal cluster-based registration method that can avoid metal artifacts during the matching procedures.
Collapse
|
36
|
Noothout JMH, De Vos BD, Wolterink JM, Postma EM, Smeets PAM, Takx RAP, Leiner T, Viergever MA, Isgum I. Deep Learning-Based Regression and Classification for Automatic Landmark Localization in Medical Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:4011-4022. [PMID: 32746142 DOI: 10.1109/tmi.2020.3009002] [Citation(s) in RCA: 35] [Impact Index Per Article: 8.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2023]
Abstract
In this study, we propose a fast and accurate method to automatically localize anatomical landmarks in medical images. We employ a global-to-local localization approach using fully convolutional neural networks (FCNNs). First, a global FCNN localizes multiple landmarks through the analysis of image patches, performing regression and classification simultaneously. In regression, displacement vectors pointing from the center of image patches towards landmark locations are determined. In classification, presence of landmarks of interest in the patch is established. Global landmark locations are obtained by averaging the predicted displacement vectors, where the contribution of each displacement vector is weighted by the posterior classification probability of the patch that it is pointing from. Subsequently, for each landmark localized with global localization, local analysis is performed. Specialized FCNNs refine the global landmark locations by analyzing local sub-images in a similar manner, i.e. by performing regression and classification simultaneously and combining the results. Evaluation was performed through localization of 8 anatomical landmarks in CCTA scans, 2 landmarks in olfactory MR scans, and 19 landmarks in cephalometric X-rays. We demonstrate that the method performs similarly to a second observer and is able to localize landmarks in a diverse set of medical images, differing in image modality, image dimensionality, and anatomical coverage.
Collapse
|
37
|
Tandel GS, Balestrieri A, Jujaray T, Khanna NN, Saba L, Suri JS. Multiclass magnetic resonance imaging brain tumor classification using artificial intelligence paradigm. Comput Biol Med 2020; 122:103804. [DOI: 10.1016/j.compbiomed.2020.103804] [Citation(s) in RCA: 46] [Impact Index Per Article: 11.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2020] [Revised: 05/01/2020] [Accepted: 05/02/2020] [Indexed: 12/18/2022]
|
38
|
Abstract
Image registration is a key pre-procedure for high level image processing. However, taking into consideration the complexity and accuracy of the algorithm, the image registration algorithm always has high time complexity. To speed up the registration algorithm, parallel computation is a relevant strategy. Parallelizing the algorithm by implementing Lattice Boltzmann method (LBM) seems a good candidate. In consequence, this paper proposes a novel parallel LBM based model (LB model) for image registration. The main idea of our method consists in simulating the convection diffusion equation through a LB model with an ad hoc collision term. By applying our method on computed tomography angiography images (CTA images), Magnet Resonance images (MR images), natural scene image and artificial images, our model proves to be faster than classical methods and achieves accurate registration. In the continuity of 2D image registration model, the LB model is extended to 3D volume registration providing excellent results in domain such as medical imaging. Our method can run on massively parallel architectures, ranging from embedded field programmable gate arrays (FPGAs) and digital signal processors (DSPs) up to graphics processing units (GPUs).
Collapse
|
39
|
Hassanin AAIM, Abd El-Samie FE, El Banby GM. A real-time approach for automatic defect detection from PCBs based on SURF features and morphological operations. MULTIMEDIA TOOLS AND APPLICATIONS 2019; 78:34437-34457. [DOI: 10.1007/s11042-019-08097-9] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/09/2018] [Revised: 07/08/2019] [Accepted: 08/13/2019] [Indexed: 09/02/2023]
|
40
|
Kubicek J, Tomanec F, Cerny M, Vilimek D, Kalova M, Oczka D. Recent Trends, Technical Concepts and Components of Computer-Assisted Orthopedic Surgery Systems: A Comprehensive Review. SENSORS (BASEL, SWITZERLAND) 2019; 19:E5199. [PMID: 31783631 PMCID: PMC6929084 DOI: 10.3390/s19235199] [Citation(s) in RCA: 23] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/20/2019] [Revised: 11/08/2019] [Accepted: 11/12/2019] [Indexed: 12/17/2022]
Abstract
Computer-assisted orthopedic surgery (CAOS) systems have become one of the most important and challenging types of system in clinical orthopedics, as they enable precise treatment of musculoskeletal diseases, employing modern clinical navigation systems and surgical tools. This paper brings a comprehensive review of recent trends and possibilities of CAOS systems. There are three types of the surgical planning systems, including: systems based on the volumetric images (computer tomography (CT), magnetic resonance imaging (MRI) or ultrasound images), further systems utilize either 2D or 3D fluoroscopic images, and the last one utilizes the kinetic information about the joints and morphological information about the target bones. This complex review is focused on three fundamental aspects of CAOS systems: their essential components, types of CAOS systems, and mechanical tools used in CAOS systems. In this review, we also outline the possibilities for using ultrasound computer-assisted orthopedic surgery (UCAOS) systems as an alternative to conventionally used CAOS systems.
Collapse
Affiliation(s)
- Jan Kubicek
- Department of Cybernetics and Biomedical Engineering, VSB-Technical University of Ostrava, FEECS, 708 00 Ostrava-Poruba, Czech Republic; (F.T.); (M.C.); (D.V.); (M.K.); (D.O.)
| | | | | | | | | | | |
Collapse
|
41
|
Abstract
The field of robotic surgery has progressed from small teams of researchers repurposing industrial robots, to a competitive and highly innovative subsection of the medical device industry. Surgical robots allow surgeons to perform tasks with greater ease, accuracy, or safety, and fall under one of four levels of autonomy; active, semi-active, passive, and remote manipulator. The increased accuracy afforded by surgical robots has allowed for cementless hip arthroplasty, improved postoperative alignment following knee arthroplasty, and reduced duration of intraoperative fluoroscopy among other benefits. Cutting of bone has historically used tools such as hand saws and drills, with other elaborate cutting tools now used routinely to remodel bone. Improvements in cutting accuracy and additional options for safety and monitoring during surgery give robotic surgeries some advantages over conventional techniques. This article aims to provide an overview of current robots and tools with a common target tissue of bone, proposes a new process for defining the level of autonomy for a surgical robot, and examines future directions in robotic surgery.
Collapse
|
42
|
Gyawali PK, Horacek BM, Sapp JL, Wang L. Sequential Factorized Autoencoder for Localizing the Origin of Ventricular Activation From 12-Lead Electrocardiograms. IEEE Trans Biomed Eng 2019; 67:1505-1516. [PMID: 31494539 DOI: 10.1109/tbme.2019.2939138] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/23/2022]
Abstract
OBJECTIVE This work presents a novel approach to handle the inter-subject variations existing in the population analysis of ECG, applied for localizing the origin of ventricular tachycardia (VT) from 12-lead electrocardiograms (ECGs). METHODS The presented method involves a factor disentangling sequential autoencoder (f-SAE) - realized in both long short-term memory (LSTM) and gated recurrent unit (GRU) networks - to learn to disentangle the inter-subject variations from the factor relating to the location of origin of VT. To perform such disentanglement, a pair-wise contrastive loss is introduced. RESULTS The presented methods are evaluated on ECG dataset with 1012 distinct pacing sites collected from scar-related VT patients during routine pace-mapping procedures. Experiments demonstrate that, for classifying the origin of VT into the predefined segments, the presented f-SAE improves the classification accuracy by 8.94% from using prescribed QRS features, by 1.5% from the supervised deep CNN network, and 5.15% from the standard SAE without factor disentanglement. Similarly, when predicting the coordinates of the VT origin, the presented f-SAE improves the performance by 2.25 mm from using prescribed QRS features, by 1.18 mm from the supervised deep CNN network and 1.6 mm from the standard SAE. CONCLUSION These results demonstrate the importance as well as the feasibility of the presented f-SAE approach for separating inter-subject variations when using 12-lead ECG to localize the origin of VT. SIGNIFICANCE This work suggests the important research direction to deal with the well-known challenge posed by inter-subject variations during population analysis from ECG signals.
Collapse
|
43
|
Sorriento A, Porfido MB, Mazzoleni S, Calvosa G, Tenucci M, Ciuti G, Dario P. Optical and Electromagnetic Tracking Systems for Biomedical Applications: A Critical Review on Potentialities and Limitations. IEEE Rev Biomed Eng 2019; 13:212-232. [PMID: 31484133 DOI: 10.1109/rbme.2019.2939091] [Citation(s) in RCA: 34] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/27/2023]
Abstract
Optical and electromagnetic tracking systems represent the two main technologies integrated into commercially-available surgical navigators for computer-assisted image-guided surgery so far. Optical Tracking Systems (OTSs) work within the optical spectrum to track the position and orientation, i.e., pose of target surgical instruments. OTSs are characterized by high accuracy and robustness to environmental conditions. The main limitation of OTSs is the need of a direct line-of-sight between the optical markers and the camera sensor, rigidly fixed into the operating theatre. Electromagnetic Tracking Systems (EMTSs) use electromagnetic field generator to detect the pose of electromagnetic sensors. EMTSs do not require such a direct line-of-sight, however the presence of metal or ferromagnetic sources in the operating workspace can significantly affect the measurement accuracy. The aim of the proposed review is to provide a complete and detailed overview of optical and electromagnetic tracking systems, including working principles, source of error and validation protocols. Moreover, commercial and research-oriented solutions, as well as clinical applications, are described for both technologies. Finally, a critical comparative analysis of the state of the art which highlights the potentialities and the limitations of each tracking system for a medical use is provided.
Collapse
|
44
|
Kmiecik B, Łabowska M, Detyna J. Determination of the difference between two complex polymer models simulating the behaviour of biological structures. Biocybern Biomed Eng 2019. [DOI: 10.1016/j.bbe.2019.04.001] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
45
|
Gomes-Fonseca J, Queirós S, Morais P, Pinho ACM, Fonseca JC, Correia-Pinto J, Lima E, Vilaça JL. Surface-based registration between CT and US for image-guided percutaneous renal access - A feasibility study. Med Phys 2019; 46:1115-1126. [DOI: 10.1002/mp.13369] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2018] [Revised: 12/13/2018] [Accepted: 12/19/2018] [Indexed: 12/30/2022] Open
Affiliation(s)
- João Gomes-Fonseca
- Life and Health Sciences Research Institute (ICVS); School of Medicine; University of Minho; Braga Portugal
- ICVS/3B's-PT; Government Associate Laboratory; Braga/Guimarães 4710-057 Portugal
| | - Sandro Queirós
- Life and Health Sciences Research Institute (ICVS); School of Medicine; University of Minho; Braga Portugal
- ICVS/3B's-PT; Government Associate Laboratory; Braga/Guimarães 4710-057 Portugal
- 2Ai; Polytechnic Institute of Cávado and Ave; Barcelos Portugal
| | - Pedro Morais
- Life and Health Sciences Research Institute (ICVS); School of Medicine; University of Minho; Braga Portugal
- ICVS/3B's-PT; Government Associate Laboratory; Braga/Guimarães 4710-057 Portugal
- 2Ai; Polytechnic Institute of Cávado and Ave; Barcelos Portugal
| | - António C. M. Pinho
- Department of Mechanical Engineering; School of Engineering; University of Minho; Guimarães Portugal
| | - Jaime C. Fonseca
- Algoritmi Center; School of Engineering; University of Minho; Guimarães Portugal
- Department of Industrial Electronics; School of Engineering; University of Minho; Guimarães Portugal
| | - Jorge Correia-Pinto
- Life and Health Sciences Research Institute (ICVS); School of Medicine; University of Minho; Braga Portugal
- ICVS/3B's-PT; Government Associate Laboratory; Braga/Guimarães 4710-057 Portugal
- Department of Pediatric Surgery; Hospital of Braga; Braga Portugal
| | - Estêvão Lima
- Life and Health Sciences Research Institute (ICVS); School of Medicine; University of Minho; Braga Portugal
- ICVS/3B's-PT; Government Associate Laboratory; Braga/Guimarães 4710-057 Portugal
- Deparment of Urology; Hospital of Braga; Braga Portugal
| | - João L. Vilaça
- Life and Health Sciences Research Institute (ICVS); School of Medicine; University of Minho; Braga Portugal
- ICVS/3B's-PT; Government Associate Laboratory; Braga/Guimarães 4710-057 Portugal
- 2Ai; Polytechnic Institute of Cávado and Ave; Barcelos Portugal
| |
Collapse
|
46
|
Reena Benjamin J, Jayasree T. Improved medical image fusion based on cascaded PCA and shift invariant wavelet transforms. Int J Comput Assist Radiol Surg 2017; 13:229-240. [PMID: 29250750 DOI: 10.1007/s11548-017-1692-4] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2017] [Accepted: 12/05/2017] [Indexed: 10/18/2022]
Abstract
PURPOSE In the medical field, radiologists need more informative and high-quality medical images to diagnose diseases. Image fusion plays a vital role in the field of biomedical image analysis. It aims to integrate the complementary information from multimodal images, producing a new composite image which is expected to be more informative for visual perception than any of the individual input images. The main objective of this paper is to improve the information, to preserve the edges and to enhance the quality of the fused image using cascaded principal component analysis (PCA) and shift invariant wavelet transforms. METHODS A novel image fusion technique based on cascaded PCA and shift invariant wavelet transforms is proposed in this paper. PCA in spatial domain extracts relevant information from the large dataset based on eigenvalue decomposition, and the wavelet transform operating in the complex domain with shift invariant properties brings out more directional and phase details of the image. The significance of maximum fusion rule applied in dual-tree complex wavelet transform domain enhances the average information and morphological details. RESULTS The input images of the human brain of two different modalities (MRI and CT) are collected from whole brain atlas data distributed by Harvard University. Both MRI and CT images are fused using cascaded PCA and shift invariant wavelet transform method. The proposed method is evaluated based on three main key factors, namely structure preservation, edge preservation, contrast preservation. The experimental results and comparison with other existing fusion methods show the superior performance of the proposed image fusion framework in terms of visual and quantitative evaluations. CONCLUSION In this paper, a complex wavelet-based image fusion has been discussed. The experimental results demonstrate that the proposed method enhances the directional features as well as fine edge details. Also, it reduces the redundant details, artifacts, distortions.
Collapse
Affiliation(s)
- J Reena Benjamin
- Electronics and Communication Engineering Department, Narayanaguru Siddhartha College of Engineering, Kanyakumari district, Tamilnadu, India.
| | - T Jayasree
- Electronics and Communication Engineering Department, Government College of Engineering, Tirunelveli, Tamilnadu, India
| |
Collapse
|