1
|
Tesema KW, Hill L, Jones MW, Ahmad MI, Tam GKL. Point Cloud Completion: A Survey. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2024; 30:6880-6899. [PMID: 38117617 DOI: 10.1109/tvcg.2023.3344935] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/22/2023]
Abstract
Point cloud completion is the task of producing a complete 3D shape given an input of a partial point cloud. It has become a vital process in 3D computer graphics, vision and applications such as autonomous driving, robotics, and augmented reality. These applications often rely on the presence of a complete 3D representation of the environment. Over the past few years, many completion algorithms have been proposed and a substantial amount of research has been carried out. However, there are not many in-depth surveys that summarise the research progress in such a way that allows users to make an informed choice of what algorithms to employ given the type of data they have, the end result they want, the challenges they may face and the possible strategies they could use. In this study, we present a comprehensive survey and classification of articles on point cloud completion untill August 2023 based on the strategies, techniques, inputs, outputs, and network architectures. We will also cover datasets, evaluation methods, and application areas in point cloud completion. Finally, we discuss challenges faced by the research community and future research directions.
Collapse
|
2
|
Tangen GA, Aadahl P, Hernes TAN, Manstad-Hulaas F. Vessel-based CTA-image to spatial anatomy registration using tracked catheter position data: preclinical evaluation of in vivo accuracy. Eur Radiol Exp 2024; 8:99. [PMID: 39196294 PMCID: PMC11358569 DOI: 10.1186/s41747-024-00499-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2024] [Accepted: 08/02/2024] [Indexed: 08/29/2024] Open
Abstract
Electromagnetic tracking of endovascular instruments has the potential to substantially decrease radiation exposure of patients and personnel. In this study, we evaluated the in vivo accuracy of a vessel-based method to register preoperative computed tomography angiography (CTA) images to physical coordinates using an electromagnetically tracked guidewire. Centerlines of the aortoiliac arteries were extracted from preoperative CTA acquired from five swine. Intravascular positions were obtained from an electromagnetically tracked guidewire. An iterative-closest-point algorithm registered the position data to the preoperative image centerlines. To evaluate the registration accuracy, a guidewire was placed inside the superior mesenteric, left and right renal arteries under fluoroscopic guidance. Position data was acquired with electromagnetic tracking as the guidewire was pulled into the aorta. The resulting measured positions were compared to the corresponding ostia manually identified in the CTA images after applying the registration. The three-dimensional (3D) Euclidean distances were calculated between each corresponding ostial point, and the root mean square (RMS) was calculated for each registration. The median 3D RMS for all registrations was 4.82 mm, with an interquartile range of 3.53-6.14 mm. A vessel-based registration of CTA images to vascular anatomy is possible with acceptable accuracy and encourages further clinical testing. RELEVANCE STATEMENT: This study shows that the centerline algorithm can be used to register preoperative CTA images to vascular anatomy, with the potential to further reduce ionizing radiation exposure during vascular procedures. KEY POINTS: Preoperative images can be used to guide the procedure without ionizing intraoperative imaging. Preoperative imaging can be the only imaging modality used for guidance of vascular procedures. No need to use external fiducial markers to register/match images and spatial anatomy. Acceptable accuracy can be achieved for navigation in a preclinical setting.
Collapse
Affiliation(s)
- Geir Arne Tangen
- SINTEF Digital, Department of Health Research, Trondheim, Norway
- Norwegian National Center for Minimally Invasive and Image-Guided Diagnostics and Therapy, St. Olavs Hospital, Trondheim, Norway
- Department of Circulation and Medical Imaging, Norwegian University of Science and Technology, Trondheim, Norway
| | - Petter Aadahl
- Department of Circulation and Medical Imaging, Norwegian University of Science and Technology, Trondheim, Norway
- Department of Cardiothoracic Anesthesia and Intensive Care, St. Olavs Hospital, Trondheim, Norway
| | - Toril A N Hernes
- Department of Circulation and Medical Imaging, Norwegian University of Science and Technology, Trondheim, Norway
| | - Frode Manstad-Hulaas
- Norwegian National Center for Minimally Invasive and Image-Guided Diagnostics and Therapy, St. Olavs Hospital, Trondheim, Norway.
- Department of Circulation and Medical Imaging, Norwegian University of Science and Technology, Trondheim, Norway.
- Future Operating Room, St. Olav University Hospital, Trondheim, Norway.
- Department of Radiology, St. Olavs Hospital, Trondheim, Norway.
| |
Collapse
|
3
|
Zhao M, Huang X, Jiang J, Mou L, Yan DM, Ma L. Accurate Registration of Cross-Modality Geometry via Consistent Clustering. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2024; 30:4055-4067. [PMID: 37027717 DOI: 10.1109/tvcg.2023.3247169] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
The registration of unitary-modality geometric data has been successfully explored over past decades. However, existing approaches typically struggle to handle cross-modality data due to the intrinsic difference between different models. To address this problem, in this article, we formulate the cross-modality registration problem as a consistent clustering process. First, we study the structure similarity between different modalities based on an adaptive fuzzy shape clustering, from which a coarse alignment is successfully operated. Then, we optimize the result using fuzzy clustering consistently, in which the source and target models are formulated as clustering memberships and centroids, respectively. This optimization casts new insight into point set registration, and substantially improves the robustness against outliers. Additionally, we investigate the effect of fuzzier in fuzzy clustering on the cross-modality registration problem, from which we theoretically prove that the classical Iterative Closest Point (ICP) algorithm is a special case of our newly defined objective function. Comprehensive experiments and analysis are conducted on both synthetic and real-world cross-modality datasets. Qualitative and quantitative results demonstrate that our method outperforms state-of-the-art approaches with higher accuracy and robustness. Our code is publicly available at https://github.com/zikai1/CrossModReg.
Collapse
|
4
|
Liu Y, Wang Z, Huang J, Zhang L. Robust Point Cloud Registration for Aircraft Engine Pipeline Systems. SENSORS (BASEL, SWITZERLAND) 2024; 24:3358. [PMID: 38894149 PMCID: PMC11175081 DOI: 10.3390/s24113358] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/26/2024] [Revised: 05/06/2024] [Accepted: 05/18/2024] [Indexed: 06/21/2024]
Abstract
Aircraft engine systems are composed of numerous pipelines. It is crucial to regularly inspect these pipelines to detect any damages or failures that could potentially lead to serious accidents. The inspection process typically involves capturing complete 3D point clouds of the pipelines using 3D scanning techniques from multiple viewpoints. To obtain a complete and accurate representation of the aircraft pipeline system, it is necessary to register and align the individual point clouds acquired from different views. However, the structures of aircraft pipelines often appear similar from different viewpoints, and the scanning process is prone to occlusions, resulting in incomplete point cloud data. The occlusions pose a challenge for existing registration methods, as they can lead to missing or wrong correspondences. To this end, we present a novel registration framework specifically designed for aircraft pipeline scenes. The proposed framework consists of two main steps. First, we extract the point feature structure of the pipeline axis by leveraging the cylindrical characteristics observed between adjacent blocks. Then, we design a new 3D descriptor called PL-PPFs (Point Line-Point Pair Features), which combines information from both the pipeline features and the engine assembly line features within the aircraft pipeline point cloud. By incorporating these relevant features, our descriptor enables accurate identification of the structure of the engine's piping system. Experimental results demonstrate the effectiveness of our approach on aircraft engine pipeline point cloud data.
Collapse
Affiliation(s)
- Yusong Liu
- College of Mechanical and Electrical Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China;
- Chengdu Aircraft Industrial (Group) Co., Ltd., Chengdu 610091, China;
| | - Zhihai Wang
- Norla Institute of Technical Physics, Chengdu 610041, China;
| | - Jichuan Huang
- Chengdu Aircraft Industrial (Group) Co., Ltd., Chengdu 610091, China;
- Systems Engineering, Northwestern Polytechnical University, Xi’an 710072, China
| | - Liyan Zhang
- College of Mechanical and Electrical Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China;
| |
Collapse
|
5
|
Elbashti M, Molinero-Mourelle P, Aswehlee A, Bornstein MM, Abou-Ayash S, Schimmel M, Ella B, Naveau A. Effect of triangular mesh resolution on the geometrical trueness of segmented CBCT maxillofacial data into STL format. J Dent 2023; 138:104722. [PMID: 37742810 DOI: 10.1016/j.jdent.2023.104722] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/27/2022] [Revised: 08/28/2023] [Accepted: 09/21/2023] [Indexed: 09/26/2023] Open
Abstract
OBJECTIVES To determine the optimal level of mesh reduction that would maintain acceptable levels of geometrical trueness while also minimizing the impact on other parameters such as file size and processing time. METHODS Intraoral and extraoral maxillofacial defects were created on 8 cadaver heads and scanned by using a CBCT scanner (NewTom 3D Imaging, Verona). DICOM data were segmented to produce head (n=8) and skull models (n=8) saved as standard tessellation language (STL) files. A further processing of head models was preformed to produce face (n=8) and ear models (n=8). A mesh reduction process was performed for each STL model (reference, R0) by generating 50% (R1), 75% (R2), and 90% (R3) reductions. The 3 datasets were compared to the R0 file using 3D evaluation software (GOM Inspect) using a global best-fit algorithm, to calculate the root mean square (RMS) deviations. Statistical analyses were performed at a level of significance of α=0.05. RESULTS There was no 3D deviation after the 50% triangular mesh reduction in the 4 datasets. Minor 3D deviations were observed after 75% reduction, in all groups. After 90% reduction, higher 3D deviations were observed, and especially in head and skull. Statistically significant increase in 3D deviations was observed with higher degrees of mesh reduction (p < 0.001). CONCLUSION The resolution of CBCT-based maxillofacial defect models can be reduced up to 50%, with neglectable concern to inaccuracy. CLINICAL SIGNIFICANCE Accurate maxillofacial models can be obtained from CBCT DICOM files after segmentation and export as STL files, even when the mesh resolution is reduced up to 50%. This information can be valuable for practitioners and researchers working with 3D models of maxillofacial defects.
Collapse
Affiliation(s)
- Mahmoud Elbashti
- Department of Maxillofacial Prosthetics, Graduate School of Medical and Dental Science, Tokyo Medical and Dental University, Tokyo, Japan.
| | - Pedro Molinero-Mourelle
- Department of Conservative Dentistry and Orofacial Prosthodontics, Complutense University of Madrid, Madrid, Spain; Department of Reconstructive Dentistry and Gerodontology, University of Bern, Bern, Switzerland.
| | - Amel Aswehlee
- Department of Dental Technology, University of Tripoli, Tripoli, Libya.
| | - Michael M Bornstein
- Department of Oral Health & Medicine, University Center for Dental Medicine Basel UZB, University of Basel, Basel, Switzerland.
| | - Samir Abou-Ayash
- Department of Reconstructive Dentistry and Gerodontology, University of Bern, Bern, Switzerland.
| | - Martin Schimmel
- Department of Reconstructive Dentistry and Gerodontology, University of Bern, Bern, Switzerland, and Division of Gerodontology and Removable Prosthodontics, University Clinics of Dental Medicine, University of Geneva, Switzerland.
| | - Bruno Ella
- Oral Surgery Department, School of Surgery, Bordeaux University Hospital, Bordeaux, France.
| | - Adrien Naveau
- Department of Prosthodontics, University of Bordeaux, Bordeaux, France.
| |
Collapse
|
6
|
Elsakka A, Park BJ, Marinelli B, Swinburne NC, Schefflein J. Virtual and Augmented Reality in Interventional Radiology: Current Applications, Challenges, and Future Directions. Tech Vasc Interv Radiol 2023; 26:100919. [PMID: 38071031 PMCID: PMC11152052 DOI: 10.1016/j.tvir.2023.100919] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/18/2023]
Abstract
Virtual reality (VR) and augmented Reality (AR) are emerging technologies with the potential to revolutionize Interventional radiology (IR). These innovations offer advantages in patient care, interventional planning, and educational training by improving the visualization and navigation of medical images. Despite progress, several challenges hinder their widespread adoption, including limitations in navigation systems, cost, clinical acceptance, and technical constraints of AR/VR equipment. However, ongoing research holds promise with recent advancements such as shape-sensing needles and improved organ deformation modeling. The development of deep learning techniques, particularly for medical imaging segmentation, presents a promising avenue to address existing accuracy and precision issues. Future applications of AR/VR in IR include simulation-based training, preprocedural planning, intraprocedural guidance, and increased patient engagement. As these technologies advance, they are expected to facilitate telemedicine, enhance operational efficiency, and improve patient outcomes, marking a new frontier in interventional radiology.
Collapse
Affiliation(s)
- Ahmed Elsakka
- Neuroradiology Service, Department of Radiology, Memorial Sloan Kettering Cancer Center, New York, NY
| | - Brian J Park
- Oregon Health & Science University, Portland, OR
| | - Brett Marinelli
- Interventional Radiology Service, Department of Radiology, Memorial Sloan Kettering Cancer Center, New York, NY
| | - Nathaniel C Swinburne
- Neuroradiology Service, Department of Radiology, Memorial Sloan Kettering Cancer Center, New York, NY
| | - Javin Schefflein
- Neuroradiology Service, Department of Radiology, Memorial Sloan Kettering Cancer Center, New York, NY.
| |
Collapse
|
7
|
Cong Y, Chen R, Ma B, Liu H, Hou D, Yang C. A Comprehensive Study of 3-D Vision-Based Robot Manipulation. IEEE TRANSACTIONS ON CYBERNETICS 2023; 53:1682-1698. [PMID: 34543212 DOI: 10.1109/tcyb.2021.3108165] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Robot manipulation, for example, pick-and-place manipulation, is broadly used for intelligent manufacturing with industrial robots, ocean engineering with underwater robots, service robots, or even healthcare with medical robots. Most traditional robot manipulations adopt 2-D vision systems with plane hypotheses and can only generate 3-DOF (degrees of freedom) pose accordingly. To mimic human intelligence and endow the robot with more flexible working capabilities, 3-D vision-based robot manipulation has been studied. However, this task is still challenging in the open world especially for general object recognition and pose estimation with occlusion in cluttered backgrounds and human-like flexible manipulation. In this article, we propose a comprehensive analysis of recent progress about the 3-D vision for robot manipulation, including 3-D data acquisition and representation, robot-vision calibration, 3-D object detection/recognition, 6-DOF pose estimation, grasping estimation, and motion planning. We then present some public datasets, evaluation criteria, comparisons, and challenges. Finally, the related application domains of robot manipulation are given, and some future directions and open problems are studied as well.
Collapse
|
8
|
Tang J, Zhou Q, Huang Y. Self-Supervised Learning for Non-Rigid Registration Between Near-Isometric 3D Surfaces in Medical Imaging. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:519-532. [PMID: 36318555 DOI: 10.1109/tmi.2022.3218662] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Non-rigid registration between 3D surfaces is an important but notorious problem in medical imaging, because finding correspondences between non-isometric instances is mathematically non-trivial. We propose a novel self-supervised method to learn shape correspondences directly from a group of bone surfaces segmented from CT scans, without any supervision from time-consuming and error-prone manual annotations. Relying on a Siamese architecture, DiffusionNet as the feature extractor is jointly trained with a pair of randomly rotated and scaled copies of the same shape. The learned embeddings are aligned in spectral domain using eigenfunctions of the Laplace-Beltrami Operator. Additional normalization and regularization losses are incorporated to guide the learned embeddings towards a similar uniform representation over spectrum, which promotes the embeddings to encode multiscale features and advocates sparsity and diagonality of the inferred functional maps. Our method achieves state-of-the-art results among the unsupervised methods on several benchmarks, and presents greater robustness and efficacy in registering moderately deformed shapes. A hybrid refinement strategy is proposed to retrieve smooth and close-to-conformal point-to-point correspondences from the inferred functional map. Our method is orientation and discretization-invariant. Given a pair of near-isometric surfaces, our method automatically computes registration in high accuracy, and outputs anatomically meaningful correspondences. In this study, we show that it is possible to use neural networks to learn general embeddings from 3D shapes in a self-supervised way. The learned features are multiscale, informative, and discriminative, which might potentially benefit almost all types of morphology-related downstream tasks, such as diagnostics, data screening and statistical shape analysis in future.
Collapse
|
9
|
Gallucci A, Znamenskiy D, Long Y, Pezzotti N, Petkovic M. Generating High-Resolution 3D Faces and Bodies Using VQ-VAE-2 with PixelSNAIL Networks on 2D Representations. SENSORS (BASEL, SWITZERLAND) 2023; 23:1168. [PMID: 36772208 PMCID: PMC9921729 DOI: 10.3390/s23031168] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/23/2022] [Revised: 01/15/2023] [Accepted: 01/16/2023] [Indexed: 06/18/2023]
Abstract
Modeling and representing 3D shapes of the human body and face is a prominent field due to its applications in the healthcare, clothes, and movie industry. In our work, we tackled the problem of 3D face and body synthesis by reducing 3D meshes to 2D image representations. We show that the face can naturally be modeled on a 2D grid. At the same time, for more challenging 3D body geometries, we proposed a novel non-bijective 3D-2D conversion method representing the 3D body mesh as a plurality of rendered projections on the 2D grid. Then, we trained a state-of-the-art vector-quantized variational autoencoder (VQ-VAE-2) to learn a latent representation of 2D images and fit a PixelSNAIL autoregressive model to sample novel synthetic meshes. We evaluated our method versus a classical one based on principal component analysis (PCA) by sampling from the empirical cumulative distribution of the PCA scores. We used the empirical distributions of two commonly used metrics, specificity and diversity, to quantitatively demonstrate that the synthetic faces generated with our method are statistically closer to real faces when compared with the PCA ones. Our experiment on the 3D body geometry requires further research to match the test set statistics but shows promising results.
Collapse
Affiliation(s)
- Alessio Gallucci
- Philips Research, 5656 AE Eindhoven, The Netherlands
- Department of Mathematics and Computer Science, Eindhoven University of Technology, 5612 AE Eindhoven, The Netherlands
| | | | - Yuxuan Long
- Philips Research, 5656 AE Eindhoven, The Netherlands
| | - Nicola Pezzotti
- Philips Research, 5656 AE Eindhoven, The Netherlands
- Department of Mathematics and Computer Science, Eindhoven University of Technology, 5612 AE Eindhoven, The Netherlands
| | - Milan Petkovic
- Philips Research, 5656 AE Eindhoven, The Netherlands
- Department of Mathematics and Computer Science, Eindhoven University of Technology, 5612 AE Eindhoven, The Netherlands
| |
Collapse
|
10
|
A method based on 3D affine alignment for the quantification of palatal expansion. PLoS One 2022; 17:e0278301. [PMID: 36584107 PMCID: PMC9803133 DOI: 10.1371/journal.pone.0278301] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2021] [Accepted: 11/14/2022] [Indexed: 12/31/2022] Open
Abstract
INTRODUCTION The current methodologies to quantify the palatal expansion are based on a preliminary rigid superimposition of 3D digital models representing the status of a given patient at different times. A new method based on affine alignment is proposed and compared to the gold standard, leading to the automatic analysis of 3-dimensional structural changes and to a simple numeric quantification of overall expansion vector and a better alignment of the digital models. MATERIALS AND METHODS 40 digital models (timing span delta 25.8 ± 12.5 months) from young patients (mean age 10.7 ± 2.6) treated with two different palatal expansion techniques (20 subjects with RME-Rapid Maxillary Expander, and 20 subjects with NiTiSE, NiTi self-expander) were superimposed with the new affine alignment technique implemented as an extension package of the open-source MeshLab, from a golden standard starting point of rigid alignment. The results were then compared. RESULTS The new measurement function indicates a mean expansion expressed in a single numeric value of 9.3%, 10.3% for the RME group and 8.4% for the NiTiSE group respectively. The comparison with the golden standard showed a decrease to the average error from 0.91 mm to 0.58 mm. CONCLUSIONS Affine alignment improves the current perspective of structural change quantification in the specific group of growing patients treated with palatal expanders giving the clinician useful information on the 3-dimensional morphological changes.
Collapse
|
11
|
Yan Z, Yi Z, Hu R, Mitra NJ, Cohen-Or D, Huang H. Consistent Two-Flow Network for Tele-Registration of Point Clouds. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2022; 28:4304-4318. [PMID: 34077360 DOI: 10.1109/tvcg.2021.3086113] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Rigid registration of partial observations is a fundamental problem in various applied fields. In computer graphics, special attention has been given to the registration between two partial point clouds generated by scanning devices. State-of-the-art registration techniques still struggle when the overlap region between the two point clouds is small, and completely fail if there is no overlap between the scan pairs. In this article, we present a learning-based technique that alleviates this problem, and allows registration between point clouds, presented in arbitrary poses, and having little or even no overlap, a setting that has been referred to as tele-registration. Our technique is based on a novel neural network design that learns a prior of a class of shapes and can complete a partial shape. The key idea is combining the registration and completion tasks in a way that reinforces each other. In particular, we simultaneously train the registration network and completion network using two coupled flows, one that register-and-complete, and one that complete-and-register, and encourage the two flows to produce a consistent result. We show that, compared with each separate flow, this two-flow training leads to robust and reliable tele-registration, and hence to a better point cloud prediction that completes the registered scans. It is also worth mentioning that each of the components in our neural network outperforms state-of-the-art methods in both completion and registration. We further analyze our network with several ablation studies and demonstrate its performance on a large number of partial point clouds, both synthetic and real-world, that have only small or no overlap.
Collapse
|
12
|
Robust two-phase registration method for three-dimensional point set under the Bayesian mixture framework. INT J MACH LEARN CYB 2022. [DOI: 10.1007/s13042-022-01673-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
13
|
Yuan X, Maharjan A. Non-rigid point set registration: recent trends and challenges. Artif Intell Rev 2022. [DOI: 10.1007/s10462-022-10292-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
|
14
|
Ferrari C, Berretti S, Pala P, Bimbo AD. A Sparse and Locally Coherent Morphable Face Model for Dense Semantic Correspondence Across Heterogeneous 3D Faces. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2022; 44:6667-6682. [PMID: 34156937 DOI: 10.1109/tpami.2021.3090942] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
The 3D Morphable Model (3DMM) is a powerful statistical tool for representing 3D face shapes. To build a 3DMM, a training set of face scans in full point-to-point correspondence is required, and its modeling capabilities directly depend on the variability contained in the training data. Thus, to increase the descriptive power of the 3DMM, establishing a dense correspondence across heterogeneous scans with sufficient diversity in terms of identities, ethnicities, or expressions becomes essential. In this manuscript, we present a fully automatic approach that leverages a 3DMM to transfer its dense semantic annotation across raw 3D faces, establishing a dense correspondence between them. We propose a novel formulation to learn a set of sparse deformation components with local support on the face that, together with an original non-rigid deformation algorithm, allow the 3DMM to precisely fit unseen faces and transfer its semantic annotation. We extensively experimented our approach, showing it can effectively generalize to highly diverse samples and accurately establish a dense correspondence even in presence of complex facial expressions. The accuracy of the dense registration is demonstrated by building a heterogeneous, large-scale 3DMM from more than 9,000 fully registered scans obtained by joining three large datasets together.
Collapse
|
15
|
Collaborative Measurement System of Dual Mobile Robots That Integrates Visual Tracking and 3D Measurement. MACHINES 2022. [DOI: 10.3390/machines10070540] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
The manufacturing accuracy of large-scale complex components determines the performance and quality of aircraft, ships, high-speed rail, and other equipment. High-precision 3D measurement plays a crucial role in ensuring manufacturing accuracy. At present, the existing measurement methods rely heavily on manual labor, which cannot satisfy the requirements of industry quality and efficiency. This paper introduces an integrated mobile robotic measurement system for the accurate and automatic 3D measurement of large-scale components with complex curved surfaces. The system consists of the mobile optical scanning measurement device, visual tracking device, and software platform, which can realize comprehensive and accurate data acquisition and stitching of large-scale complex components. The combination of visual tracking and 3D measurement based on the coordinated motion of the dual robot achieved the automatic data acquisition of large-scale complex components without sticking coded targets. Additionally, this paper also introduces a DeepMerge algorithm that combines local and global features of the point cloud, which effectively corrects the initial stitching error of the visual tracking system. The validity of the measurement system and method was shown by the measurement and stitching experiments on the surface of the vehicle nose, ensuring the accurate measurement of the robot’s wide range of motion.
Collapse
|
16
|
Fast and robust active camera relocalization in the wild for fine-grained change detection. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.04.102] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
17
|
Zhang J, Yao Y, Deng B. Fast and Robust Iterative Closest Point. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2022; 44:3450-3466. [PMID: 33497327 DOI: 10.1109/tpami.2021.3054619] [Citation(s) in RCA: 18] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
The iterative closest point (ICP) algorithm and its variants are a fundamental technique for rigid registration between two point sets, with wide applications in different areas from robotics to 3D reconstruction. The main drawbacks for ICP are its slow convergence, as well as its sensitivity to outliers, missing data, and partial overlaps. Recent work such as Sparse ICP achieves robustness via sparsity optimization at the cost of computational speed. In this paper, we propose a new method for robust registration with fast convergence. First, we show that the classical point-to-point ICP can be treated as a majorization-minimization (MM) algorithm, and propose an Anderson acceleration approach to speed up its convergence. In addition, we introduce a robust error metric based on the Welsch's function, which is minimized efficiently using the MM algorithm with Anderson acceleration. On challenging datasets with noises and partial overlaps, we achieve similar or better accuracy than Sparse ICP while being at least an order of magnitude faster. Finally, we extend the robust formulation to point-to-plane ICP, and solve the resulting problem using a similar Anderson-accelerated MM strategy. Our robust ICP methods improve the registration accuracy on benchmark datasets while being competitive in computational time.
Collapse
|
18
|
A 3D Image Registration Method for Laparoscopic Liver Surgery Navigation. ELECTRONICS 2022. [DOI: 10.3390/electronics11111670] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/01/2023]
Abstract
At present, laparoscopic augmented reality (AR) navigation has been applied to minimally invasive abdominal surgery, which can help doctors to see the location of blood vessels and tumors in organs, so as to perform precise surgery operations. Image registration is the process of optimally mapping one or more images to the target image, and it is also the core of laparoscopic AR navigation. The key is how to shorten the registration time and optimize the registration accuracy. We have studied the three-dimensional (3D) image registration technology in laparoscopic liver surgery navigation and proposed a new registration method combining rough registration and fine registration. First, the adaptive fireworks algorithm (AFWA) is applied to rough registration, and then the optimized iterative closest point (ICP) algorithm is applied to fine registration. We proposed a method that is validated by the computed tomography (CT) dataset 3D-IRCADb-01. Experimental results show that our method is superior to other registration methods based on stochastic optimization algorithms in terms of registration time and accuracy.
Collapse
|
19
|
Chen Y, He K, Hao B, Weng Y, Chen Z. FractureNet: A 3D Convolutional Neural Network Based on the Architecture of m-Ary Tree for Fracture Type Identification. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:1196-1207. [PMID: 34890325 DOI: 10.1109/tmi.2021.3134650] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
To address the problem of automatic identification of fine-grained fracture types, in this paper, we propose a novel framework using 3D convolutional neural network (CNN) to learn fracture features from voxelized bone models which are obtained by establishing isomorphic mapping from fractured bones to a voxelized template. The network, which is named FractureNet, consists of four discriminators forming a multi-stage hierarchy. Each discriminator includes multiple sub-classifiers. These sub-classifiers are chained by two kinds of feature chains (feature map chain and classification feature chain) in the form of a full m-ary tree to perform multi-stage classification tasks. The features learned and classification results obtained at previous stages serve as prior knowledge for current learning and classification. All sub-classifiers are jointly learned in an end-to-end network via a multi-stage loss function integrating losses of the four discriminators. To make our FractureNet more robust and accurate, a data augmentation strategy termed r-combination with constraints is further proposed on the basis of an adjacency relation and a continuity relation between voxels to create a large-scale fracture dataset of voxel models. Extensive experiments show that the proposed method can recognize various fracture types in patients accurately and effectively, and enables significant improvements over the state-of-the-arts on a variety of fracture recognition tasks. Moreover, ancillary experiments on the CIFAR-10 and the PadChest datasets at large scales further support the superior performance of the proposed FractureNet.
Collapse
|
20
|
Camera, LiDAR and Multi-modal SLAM Systems for Autonomous Ground Vehicles: a Survey. J INTELL ROBOT SYST 2022. [DOI: 10.1007/s10846-022-01582-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
21
|
Abstract
Building information modelling (BIM) is evolving significantly in the architecture, engineering and construction industries. BIM involves various remote-sensing tools, procedures and standards that are useful for collating the semantic information required to produce 3D models. This is thanks to LiDAR technology, which has become one of the key elements in BIM, useful to capture a semantically rich geometric representation of 3D models in terms of 3D point clouds. This review paper explains the ‘Scan to BIM’ methodology in detail. The paper starts by summarising the 3D point clouds of LiDAR and photogrammetry. LiDAR systems based on different platforms, such as mobile, terrestrial, spaceborne and airborne, are outlined and compared. In addition, the importance of integrating multisource data is briefly discussed. Various methodologies involved in point-cloud processing such as sampling, registration and semantic segmentation are explained in detail. Furthermore, different open BIM standards are summarised and compared. Finally, current limitations and future directions are highlighted to provide useful solutions for efficient BIM models.
Collapse
|
22
|
Fast Registration of Terrestrial LiDAR Point Clouds Based on Gaussian-Weighting Projected Image Matching. REMOTE SENSING 2022. [DOI: 10.3390/rs14061466] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/07/2022]
Abstract
Terrestrial point cloud registration plays an important role in 3D reconstruction, heritage restoration and topographic mapping, etc. Unfortunately, current research studies heavily rely on matching the 3D features of overlapped areas between point clouds, which is error-prone and time-consuming. To this end, we propose an automatic point cloud registration method based on Gaussian-weighting projected image matching, which can quickly and robustly register multi-station terrestrial point clouds. Firstly, the point cloud is regularized into a 2D grid, and the point density of each cell in the grid is normalized by our Gaussian-weighting function. A grayscale image is subsequently generated by shifting and scaling the x-y coordinates of the grid to the image coordinates. Secondly, the scale-invariant features (SIFT) algorithm is used to perform image matching, and a line segment endpoint verification method is proposed to filter out negative matches. Thirdly, the transformation matrix between point clouds from two adjacent stations is calculated based on reliable image matching. Finally, a global least-square optimization is conducted to align multi-station point clouds and then obtain a complete model. To test the performance of our framework, we carry out the experiment on six datasets. Compared to previous work, our method achieves the state-of-the-art performance on both efficiency and accuracy. In terms of efficiency, our method is comparable to an existing projection-based methods and 4 times faster on the indoor datasets and 10 times faster on the outdoor datasets than 4PCS-based methods. In terms of accuracy, our framework is ~2 times better than the existing projection-based method and 6 times better than 4PCS-based methods.
Collapse
|
23
|
IMU-Aided Registration of MLS Point Clouds Using Inertial Trajectory Error Model and Least Squares Optimization. REMOTE SENSING 2022. [DOI: 10.3390/rs14061365] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/27/2023]
Abstract
Mobile laser scanning (MLS) point cloud registration plays a critical role in mobile 3D mapping and inspection, but conventional point cloud registration methods for terrain LiDAR scanning (TLS) are not suitable for MLS. To cope with this challenge, we use inertial measurement unit (IMU) to assist registration and propose an MLS point cloud registration method based on an inertial trajectory error model. First, we propose an error model of inertial trajectory over a short time period to construct the constraints between trajectory points at different times. On this basis, a relationship between the point cloud registration error and the inertial trajectory error is established, then trajectory error parameters are estimated by minimizing the point cloud registration error using the least squares optimization. Finally, a reliable and concise inertial-assisted MLS registration algorithm is realized. We carried out experiments in three different scenarios: indoor, outdoor and integrated indoor–outdoor. We evaluated the overall performance, accuracy and efficiency of the proposed method. Compared with the ICP method, the accuracy and speed of the proposed method were improved by 2 and 2.8 times, respectively, which verified the effectiveness and reliability of the proposed method. Furthermore, experimental results show the significance of our method in constructing a reliable and scalable mobile 3D mapping system suitable for complex scenes.
Collapse
|
24
|
A Robust Rigid Registration Framework of 3D Indoor Scene Point Clouds Based on RGB-D Information. REMOTE SENSING 2021. [DOI: 10.3390/rs13234755] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Rigid registration of 3D indoor scenes is a fundamental yet vital task in various fields that include remote sensing (e.g., 3D reconstruction of indoor scenes), photogrammetry measurement, geometry modeling, etc. Nevertheless, state-of-the-art registration approaches still have defects when dealing with low-quality indoor scene point clouds derived from consumer-grade RGB-D sensors. The major challenge is accurately extracting correspondences between a pair of low-quality point clouds when they contain considerable noise, outliers, or weak texture features. To solve the problem, we present a point cloud registration framework in view of RGB-D information. First, we propose a point normal filter for effectively removing noise and simultaneously maintaining sharp geometric features and smooth transition regions. Second, we design a correspondence extraction scheme based on a novel descriptor encoding textural and geometry information, which can robustly establish dense correspondences between a pair of low-quality point clouds. Finally, we propose a point-to-plane registration technology via a nonconvex regularizer, which can further diminish the influence of those false correspondences and produce an exact rigid transformation between a pair of point clouds. Compared to existing state-of-the-art techniques, intensive experimental results demonstrate that our registration framework is excellent visually and numerically, especially for dealing with low-quality indoor scenes.
Collapse
|
25
|
Cho J, Kim YJ, Sunwoo L, Lee GP, Nguyen TQ, Cho SJ, Baik SH, Bae YJ, Choi BS, Jung C, Sohn CH, Han JH, Kim CY, Kim KG, Kim JH. Deep Learning-Based Computer-Aided Detection System for Automated Treatment Response Assessment of Brain Metastases on 3D MRI. Front Oncol 2021; 11:739639. [PMID: 34778056 PMCID: PMC8579083 DOI: 10.3389/fonc.2021.739639] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2021] [Accepted: 09/30/2021] [Indexed: 11/13/2022] Open
Abstract
BACKGROUND Although accurate treatment response assessment for brain metastases (BMs) is crucial, it is highly labor intensive. This retrospective study aimed to develop a computer-aided detection (CAD) system for automated BM detection and treatment response evaluation using deep learning. METHODS We included 214 consecutive MRI examinations of 147 patients with BM obtained between January 2015 and August 2016. These were divided into the training (174 MR images from 127 patients) and test datasets according to temporal separation (temporal test set #1; 40 MR images from 20 patients). For external validation, 24 patients with BM and 11 patients without BM from other institutions were included (geographic test set). In addition, we included 12 MRIs from BM patients obtained between August 2017 and March 2020 (temporal test set #2). Detection sensitivity, dice similarity coefficient (DSC) for segmentation, and agreements in one-dimensional and volumetric Response Assessment in Neuro-Oncology Brain Metastases (RANO-BM) criteria between CAD and radiologists were assessed. RESULTS In the temporal test set #1, the sensitivity was 75.1% (95% confidence interval [CI]: 69.6%, 79.9%), mean DSC was 0.69 ± 0.22, and false-positive (FP) rate per scan was 0.8 for BM ≥ 5 mm. Agreements in the RANO-BM criteria were moderate (κ, 0.52) and substantial (κ, 0.68) for one-dimensional and volumetric, respectively. In the geographic test set, sensitivity was 87.7% (95% CI: 77.2%, 94.5%), mean DSC was 0.68 ± 0.20, and FP rate per scan was 1.9 for BM ≥ 5 mm. In the temporal test set #2, sensitivity was 94.7% (95% CI: 74.0%, 99.9%), mean DSC was 0.82 ± 0.20, and FP per scan was 0.5 (6/12) for BM ≥ 5 mm. CONCLUSIONS Our CAD showed potential for automated treatment response assessment of BM ≥ 5 mm.
Collapse
Affiliation(s)
- Jungheum Cho
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam, South Korea
| | - Young Jae Kim
- Department of Biomedical Engineering, Gachon University Gil Medical Center, Incheon, South Korea
| | - Leonard Sunwoo
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam, South Korea
- Center for Artificial Intelligence in Healthcare, Seoul National University Bundang Hospital, Seongnam, South Korea
| | - Gi Pyo Lee
- Department of Biomedical Engineering, Gachon University Gil Medical Center, Incheon, South Korea
| | - Toan Quang Nguyen
- Department of Radiology, Vietnam National Cancer Hospital, Hanoi, Vietnam
| | - Se Jin Cho
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam, South Korea
| | - Sung Hyun Baik
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam, South Korea
| | - Yun Jung Bae
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam, South Korea
| | - Byung Se Choi
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam, South Korea
| | - Cheolkyu Jung
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam, South Korea
| | - Chul-Ho Sohn
- Department of Radiology, Seoul National University Hospital, Seoul, South Korea
| | - Jung-Ho Han
- Department of Neurosurgery, Seoul National University Bundang Hospital, Seongnam, South Korea
| | - Chae-Yong Kim
- Department of Neurosurgery, Seoul National University Bundang Hospital, Seongnam, South Korea
| | - Kwang Gi Kim
- Department of Biomedical Engineering, Gachon University Gil Medical Center, Incheon, South Korea
| | - Jae Hyoung Kim
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam, South Korea
| |
Collapse
|
26
|
A Fast and Robust Rotation Search and Point Cloud Registration Method for 2D Stitching and 3D Object Localization. APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app11209775] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Rotation search and point cloud registration are two fundamental problems in robotics, geometric vision, and remote sensing, which aim to estimate the rotation and transformation between the 3D vector sets and point clouds, respectively. Due to the presence of outliers (probably in very large numbers) among the putative vector or point correspondences in real-world applications, robust estimation is of great importance. In this paper, we present Inlier searching using COmpatible Structures (ICOS), a novel, efficient, and highly robust solver for both the correspondence-based rotation search and point cloud registration problems. Specifically, we (i) propose and construct a series of compatible structures for the two problems, based on which various invariants can be established, and (ii) design time-efficient frameworks to filter out outliers and seek inliers from the invariant-constrained random sampling based on the compatible structures proposed. In this manner, even with extreme outlier ratios, inliers can be effectively sifted out and collected for solving the optimal rotation and transformation, leading to our robust solver ICOS. Through plentiful experiments over standard datasets, we demonstrated that: (i) our solver ICOS is fast, accurate, and robust against over 95% outliers with nearly a 100% recall ratio of inliers for rotation search and both known-scale and unknown-scale registration, outperforming other state-of-the-art methods, and (ii) ICOS is practical for use in real-world application problems including 2D image stitching and 3D object localization.
Collapse
|
27
|
Min Z, Liu J, Liu L, Meng MQH. Generalized Coherent Point Drift With Multi-Variate Gaussian Distribution and Watson Distribution. IEEE Robot Autom Lett 2021. [DOI: 10.1109/lra.2021.3093011] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
28
|
Liao Q, Sun D, Andreasson H. Point Set Registration for 3D Range Scans Using Fuzzy Cluster-Based Metric and Efficient Global Optimization. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2021; 43:3229-3246. [PMID: 32149624 DOI: 10.1109/tpami.2020.2978477] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
This study presents a new point set registration method to align 3D range scans. In our method, fuzzy clusters are utilized to represent a scan, and the registration of two given scans is realized by minimizing a fuzzy weighted sum of the distances between their fuzzy cluster centers. This fuzzy cluster-based metric has a broad basin of convergence and is robust to noise. Moreover, this metric provides analytic gradients, allowing standard gradient-based algorithms to be applied for optimization. Based on this metric, the outlier issues are addressed. In addition, for the first time in rigid point set registration, a registration quality assessment in the absence of ground truth is provided. Furthermore, given specified rotation and translation spaces, we derive the upper and lower bounds of the fuzzy cluster-based metric and develop a branch-and-bound (BnB)-based optimization scheme, which can globally minimize the metric regardless of the initialization. This optimization scheme is performed in an efficient coarse-to-fine fashion: First, fuzzy clustering is applied to describe each of the two given scans by a small number of fuzzy clusters. Then, a global search, which integrates BnB and gradient-based algorithms, is implemented to achieve a coarse alignment for the two scans. During the global search, the registration quality assessment offers a beneficial stop criterion to detect whether a good result is obtained. Afterwards, a relatively large number of points of the two scans are directly taken as the fuzzy cluster centers, and then, the coarse solution is refined to be an exact alignment using the gradient-based local convergence. Compared to existing counterparts, this optimization scheme makes a large improvement in terms of robustness and efficiency by virtue of the fuzzy cluster-based metric and the registration quality assessment. In the experiments, the registration results of several 3D range scan pairs demonstrate the accuracy and effectiveness of the proposed method, as well as its superiority to state-of-the-art registration approaches.
Collapse
|
29
|
Abstract
AbstractStandard registration algorithms need to be independently applied to each surface to register, following careful pre-processing and hand-tuning. Recently, learning-based approaches have emerged that reduce the registration of new scans to running inference with a previously-trained model. The potential benefits are multifold: inference is typically orders of magnitude faster than solving a new instance of a difficult optimization problem, deep learning models can be made robust to noise and corruption, and the trained model may be re-used for other tasks, e.g. through transfer learning. In this paper, we cast the registration task as a surface-to-surface translation problem, and design a model to reliably capture the latent geometric information directly from raw 3D face scans. We introduce Shape-My-Face (SMF), a powerful encoder-decoder architecture based on an improved point cloud encoder, a novel visual attention mechanism, graph convolutional decoders with skip connections, and a specialized mouth model that we smoothly integrate with the mesh convolutions. Compared to the previous state-of-the-art learning algorithms for non-rigid registration of face scans, SMF only requires the raw data to be rigidly aligned (with scaling) with a pre-defined face template. Additionally, our model provides topologically-sound meshes with minimal supervision, offers faster training time, has orders of magnitude fewer trainable parameters, is more robust to noise, and can generalize to previously unseen datasets. We extensively evaluate the quality of our registrations on diverse data. We demonstrate the robustness and generalizability of our model with in-the-wild face scans across different modalities, sensor types, and resolutions. Finally, we show that, by learning to register scans, SMF produces a hybrid linear and non-linear morphable model. Manipulation of the latent space of SMF allows for shape generation, and morphing applications such as expression transfer in-the-wild. We train SMF on a dataset of human faces comprising 9 large-scale databases on commodity hardware.
Collapse
|
30
|
Andreß S, Achilles F, Bischoff J, Kußmaul AC, Böcker W, Weidert S. A method for finding high accuracy surface zones on 3D printed bone models. Comput Biol Med 2021; 135:104590. [PMID: 34216887 DOI: 10.1016/j.compbiomed.2021.104590] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2021] [Revised: 06/07/2021] [Accepted: 06/16/2021] [Indexed: 11/26/2022]
Abstract
The use of three-dimensional (3D) printing for surgical applications is steadily increasing. Errors in the printed models can lead to complications, especially when the model is used for surgery planning or diagnostics. In patient care, the validation of printed models should therefore be performed routinely. However, there currently is no standard method to determine whether the printed model meets the necessary quality requirements. In this work, we present a method that not only finds surface deviations of a printed model, but also shows high accuracy zones of a potentially corrupted model, that are safe to be used for surgery planning. Our method was tested on printed patient bone models with acetabular fractures and was compared to two common methods in orthopedics, simple landmark registration as well as landmark plus subsequent iterative closest point registration. In order to find suitable parameters and to evaluate the performance of our method, 15 digital acetabular bone models were artificially deformed, imitating four typical 3D printing errors. A sensitivity of over 95% and a specificity of over 99% was observed in finding these surface deformations. Then, the method was applied to 32 printed models that had been re-digitized using a computed tomography scanner. It was found that only 25% of these printed models were free of significant deformations. However, focussing on two common implant locations, our method revealed that 72% of the models were within the acceptable error tolerance. In comparison, simple landmark registration resulted in a 9% acceptance rate and landmark registration followed by iterative closest point registration resulted in a 41% acceptance rate. This outcome shows that our method, named Similarity Subgroups Registration, allows clinicians to safely use partially corrupted 3D printed models for surgery planning. This improves efficiency and reduces time to treatment by avoiding reprints. The similarity subgroups registration is applicable in further clinical domains as well as non-medical applications that share the requirement of local high accuracy zones on the surface of a 3D model.
Collapse
Affiliation(s)
- Sebastian Andreß
- Department of General, Trauma and Reconstructive Surgery, University Hospital, LMU Munich, Germany.
| | - Felix Achilles
- Department of General, Trauma and Reconstructive Surgery, University Hospital, LMU Munich, Germany
| | - Jonathan Bischoff
- Department of General, Trauma and Reconstructive Surgery, University Hospital, LMU Munich, Germany
| | | | - Wolfgang Böcker
- Department of General, Trauma and Reconstructive Surgery, University Hospital, LMU Munich, Germany
| | - Simon Weidert
- Department of General, Trauma and Reconstructive Surgery, University Hospital, LMU Munich, Germany
| |
Collapse
|
31
|
RegARD: Symmetry-Based Coarse Registration of Smartphone’s Colorful Point Clouds with CAD Drawings for Low-Cost Digital Twin Buildings. REMOTE SENSING 2021. [DOI: 10.3390/rs13101882] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Coarse registration of 3D point clouds plays an indispensable role for parametric, semantically rich, and realistic digital twin buildings (DTBs) in the practice of GIScience, manufacturing, robotics, architecture, engineering, and construction. However, the existing methods have prominently been challenged by (i) the high cost of data collection for numerous existing buildings and (ii) the computational complexity from self-similar layout patterns. This paper studies the registration of two low-cost data sets, i.e., colorful 3D point clouds captured by smartphones and 2D CAD drawings, for resolving the first challenge. We propose a novel method named ‘Registration based on Architectural Reflection Detection’ (RegARD) for transforming the self-symmetries in the second challenge from a barrier of coarse registration to a facilitator. First, RegARD detects the innate architectural reflection symmetries to constrain the rotations and reduce degrees of freedom. Then, a nonlinear optimization formulation together with advanced optimization algorithms can overcome the second challenge. As a result, high-quality coarse registration and subsequent low-cost DTBs can be created with semantic components and realistic appearances. Experiments showed that the proposed method outperformed existing methods considerably in both effectiveness and efficiency, i.e., 49.88% less error and 73.13% less time, on average. The RegARD presented in this paper first contributes to coarse registration theories and exploitation of symmetries and textures in 3D point clouds and 2D CAD drawings. For practitioners in the industries, RegARD offers a new automatic solution to utilize ubiquitous smartphone sensors for massive low-cost DTBs.
Collapse
|
32
|
A Strip Adjustment Method of UAV-Borne LiDAR Point Cloud Based on DEM Features for Mountainous Area. SENSORS 2021; 21:s21082782. [PMID: 33920866 PMCID: PMC8071264 DOI: 10.3390/s21082782] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/11/2021] [Revised: 04/03/2021] [Accepted: 04/12/2021] [Indexed: 11/19/2022]
Abstract
Due to the trajectory error of the low-precision position and orientation system (POS) used in unmanned aerial laser scanning (ULS), discrepancies usually exist between adjacent LiDAR (Light Detection and Ranging) strips. Strip adjustment is an effective way to eliminate these discrepancies. However, it is difficult to apply existing strip adjustment methods in mountainous areas with few artificial objects. Thus, digital elevation model-iterative closest point (DEM-ICP), a pair-wise registration method that takes topography features into account, is proposed in this paper. First, DEM-ICP filters the point clouds to remove the non-ground points. Second, the ground points are interpolated to generate continuous DEMs. Finally, a point-to-plane ICP algorithm is performed to register the adjacent DEMs with the overlapping area. A graph-based optimization is utilized following DEM-ICP to estimate the correction parameters and achieve global consistency between all strips. Experiments were carried out using eight strips collected by ULS in mountainous areas to evaluate the proposed method. The average root-mean-square error (RMSE) of all data was less than 0.4 m after the proposed strip adjustment, which was only 0.015 m higher than the result of manual registration (ground truth). In addition, the plane fitting accuracy of lateral point clouds was improved 4.2-fold, from 1.565 to 0.375 m, demonstrating the robustness and accuracy of the proposed method.
Collapse
|
33
|
Affiliation(s)
- Heng Yang
- Laboratory for Information & Decision Systems (LIDS), Massachusetts Institute of Technology, Cambridge, MA, USA
| | - Jingnan Shi
- Laboratory for Information & Decision Systems (LIDS), Massachusetts Institute of Technology, Cambridge, MA, USA
| | - Luca Carlone
- Laboratory for Information & Decision Systems (LIDS), Massachusetts Institute of Technology, Cambridge, MA, USA
| |
Collapse
|
34
|
Wu R, Yu Z, Ding D, Lu Q, Pan Z, Li H. OICP: An Online Fast Registration Algorithm Based on Rigid Translation Applied to Wire Arc Additive Manufacturing of Mold Repair. MATERIALS 2021; 14:ma14061563. [PMID: 33810152 PMCID: PMC8004783 DOI: 10.3390/ma14061563] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/10/2021] [Revised: 03/10/2021] [Accepted: 03/17/2021] [Indexed: 11/17/2022]
Abstract
As promising technology with low requirements and high depositing efficiency, Wire Arc Additive Manufacturing (WAAM) can significantly reduce the repair cost and improve the formation quality of molds. To further improve the accuracy of WAAM in repairing molds, the point cloud model that expresses the spatial distribution and surface characteristics of the mold is proposed. Since the mold has a large size, it is necessary to be scanned multiple times, resulting in multiple point cloud models. The point cloud registration, such as the Iterative Closest Point (ICP) algorithm, then plays the role of merging multiple point cloud models to reconstruct a complete data model. However, using the ICP algorithm to merge large point clouds with a low-overlap area is inefficient, time-consuming, and unsatisfactory. Therefore, this paper provides the improved Offset Iterative Closest Point (OICP) algorithm, which is an online fast registration algorithm suitable for intelligent WAAM mold repair technology. The practicality and reliability of the algorithm are illustrated by the comparison results with the standard ICP algorithm and the three-coordinate measuring instrument in the Experimental Setup Section. The results are that the OICP algorithm is feasible for registrations with low overlap rates. For an overlap rate lower than 60% in our experiments, the traditional ICP algorithm failed, while the Root Mean Square (RMS) error reached 0.1 mm, and the rotation error was within 0.5 degrees, indicating the improvement of the proposed OICP algorithm.
Collapse
Affiliation(s)
- Ruibing Wu
- School of Mechatronic Engineering and Automation, Foshan University, Foshan 528225, China; (R.W.); (Z.Y.); (Q.L.)
| | - Ziping Yu
- School of Mechatronic Engineering and Automation, Foshan University, Foshan 528225, China; (R.W.); (Z.Y.); (Q.L.)
- School of Mechanical, Materials, and Mechatronics Engineering, Faculty of Engineering and Information Sciences, University of Wollongong, Northfield Ave., Wollongong, NSW 2500, Australia; (Z.P.); (H.L.)
| | - Donghong Ding
- School of Mechatronic Engineering and Automation, Foshan University, Foshan 528225, China; (R.W.); (Z.Y.); (Q.L.)
- Correspondence:
| | - Qinghua Lu
- School of Mechatronic Engineering and Automation, Foshan University, Foshan 528225, China; (R.W.); (Z.Y.); (Q.L.)
| | - Zengxi Pan
- School of Mechanical, Materials, and Mechatronics Engineering, Faculty of Engineering and Information Sciences, University of Wollongong, Northfield Ave., Wollongong, NSW 2500, Australia; (Z.P.); (H.L.)
| | - Huijun Li
- School of Mechanical, Materials, and Mechatronics Engineering, Faculty of Engineering and Information Sciences, University of Wollongong, Northfield Ave., Wollongong, NSW 2500, Australia; (Z.P.); (H.L.)
| |
Collapse
|
35
|
Navaei Lavasani S, Farnia P, Najafzadeh E, Saghatchi S, Samavati M, Abtahi H, Deevband M, Ahmadian A. Bronchoscope motion tracking using centerline-guided Gaussian mixture model in navigated bronchoscopy. Phys Med Biol 2021; 66:025001. [PMID: 33181494 DOI: 10.1088/1361-6560/abca07] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/24/2022]
Abstract
Electromagnetic-based navigation bronchoscopy requires accurate and robust estimation of the bronchoscope position inside the bronchial tree. However, respiratory motion, coughing, patient movement, and airway deformation inflicted by bronchoscope significantly hinder the accuracy of intraoperative bronchoscopic localization. In this study, a real-time and automatic registration procedure was proposed to superimpose the current location of the bronchoscope to corresponding locations on a centerline extracted from bronchial computed tomography (CT) images. A centerline-guided Gaussian mixture model (CG-GMM) was introduced to register a bronchoscope's position concerning extracted centerlines. A GMM was fitted to bronchoscope positions where the orientation likelihood was chosen to assign the membership probabilities of the mixture model, which led to preserving the global and local structures. The problem was formulated and solved under the expectation maximization framework, where the feature correspondence and spatial transformation are estimated iteratively. Validation was performed on a dynamic phantom with four different respiratory motions and four human real bronchoscopy (RB) datasets. Results of the experiments conducted on the bronchial phantom showed that the average positional tracking error using the proposed approach was equal to 1.98 [Formula: see text] 0.98 mm that was reduced in comparison with independent electromagnetic tracking (EMT), iterative closest point (ICP), and coherent point drift (CPD) methods by 64%, 58%, and 53%, respectively. In the patient assessment part of the study, the average positional tracking error was 4.73 [Formula: see text] 4.76 mm and compared to ICP, and CPD methods showed 31.4% improvement of successfully tracked frames. Our approach introduces a novel method for real-time respiratory motion compensation that provides reliable guidance during bronchoscopic interventions and, thus could increase the diagnostic yield of transbronchial biopsy.
Collapse
Affiliation(s)
- Saeedeh Navaei Lavasani
- Biomedical Engineering and Medical Physics Department, School of Medicine, Shahid Beheshti University of Medical Sciences (SBMU), Tehran, Iran. Research Centre of Biomedical Technology and Robotics (RCBTR), Imam Khomeini Hospital Complex, Tehran University of Medical Sciences, Tehran, Iran
| | | | | | | | | | | | | | | |
Collapse
|
36
|
Shabayek AER, Aouada D. Dense and Sparse 3D Deformation Signatures for 3D Dynamic Face Recognition. IEEE ACCESS 2021; 9:38687-38705. [DOI: 10.1109/access.2021.3064179] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/02/2023]
|
37
|
Chung M, Lee J, Song W, Song Y, Yang IH, Lee J, Shin YG. Automatic Registration Between Dental Cone-Beam CT and Scanned Surface via Deep Pose Regression Neural Networks and Clustered Similarities. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:3900-3909. [PMID: 32746134 DOI: 10.1109/tmi.2020.3007520] [Citation(s) in RCA: 17] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Computerized registration between maxillofacial cone-beam computed tomography (CT) images and a scanned dental model is an essential prerequisite for surgical planning for dental implants or orthognathic surgery. We propose a novel method that performs fully automatic registration between a cone-beam CT image and an optically scanned model. To build a robust and automatic initial registration method, deep pose regression neural networks are applied in a reduced domain (i.e., two-dimensional image). Subsequently, fine registration is performed using optimal clusters. A majority voting system achieves globally optimal transformations while each cluster attempts to optimize local transformation parameters. The coherency of clusters determines their candidacy for the optimal cluster set. The outlying regions in the iso-surface are effectively removed based on the consensus among the optimal clusters. The accuracy of registration is evaluated based on the Euclidean distance of 10 landmarks on a scanned model, which have been annotated by experts in the field. The experiments show that the registration accuracy of the proposed method, measured based on the landmark distance, outperforms the best performing existing method by 33.09%. In addition to achieving high accuracy, our proposed method neither requires human interactions nor priors (e.g., iso-surface extraction). The primary significance of our study is twofold: 1) the employment of lightweight neural networks, which indicates the applicability of neural networks in extracting pose cues that can be easily obtained and 2) the introduction of an optimal cluster-based registration method that can avoid metal artifacts during the matching procedures.
Collapse
|
38
|
|
39
|
When Deep Learning Meets Data Alignment: A Review on Deep Registration Networks (DRNs). APPLIED SCIENCES-BASEL 2020. [DOI: 10.3390/app10217524] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/25/2023]
Abstract
This paper reviews recent deep learning-based registration methods. Registration is the process that computes the transformation that aligns datasets, and the accuracy of the result depends on multiple factors. The most significant factors are the size of input data; the presence of noise, outliers and occlusions; the quality of the extracted features; real-time requirements; and the type of transformation, especially those defined by multiple parameters, such as non-rigid deformations. Deep Registration Networks (DRNs) are those architectures trying to solve the alignment task using a learning algorithm. In this review, we classify these methods according to a proposed framework based on the traditional registration pipeline. This pipeline consists of four steps: target selection, feature extraction, feature matching, and transform computation for the alignment. This new paradigm introduces a higher-level understanding of registration, which makes explicit the challenging problems of traditional approaches. The main contribution of this work is to provide a comprehensive starting point to address registration problems from a learning-based perspective and to understand the new range of possibilities.
Collapse
|
40
|
Brouwer CL, Boukerroui D, Oliveira J, Looney P, Steenbakkers RJ, Langendijk JA, Both S, Gooding MJ. Assessment of manual adjustment performed in clinical practice following deep learning contouring for head and neck organs at risk in radiotherapy. Phys Imaging Radiat Oncol 2020; 16:54-60. [PMID: 33458344 PMCID: PMC7807591 DOI: 10.1016/j.phro.2020.10.001] [Citation(s) in RCA: 22] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2020] [Revised: 09/30/2020] [Accepted: 10/01/2020] [Indexed: 12/30/2022] Open
Abstract
BACKGROUND AND PURPOSE Auto-contouring performance has been widely studied in development and commissioning studies in radiotherapy, and its impact on clinical workflow assessed in that context. This study aimed to evaluate the manual adjustment of auto-contouring in routine clinical practice and to identify improvements regarding the auto-contouring model and clinical user interaction, to improve the efficiency of auto-contouring. MATERIALS AND METHODS A total of 103 clinical head and neck cancer cases, contoured using a commercial deep-learning contouring system and subsequently checked and edited for clinical use were retrospectively taken from clinical data over a twelve-month period (April 2019-April 2020). The amount of adjustment performed was calculated, and all cases were registered to a common reference frame for assessment purposes. The median, 10th and 90th percentile of adjustment were calculated and displayed using 3D renderings of structures to visually assess systematic and random adjustment. Results were also compared to inter-observer variation reported previously. Assessment was performed for both the whole structures and for regional sub-structures, and according to the radiation therapy technologist (RTT) who edited the contour. RESULTS The median amount of adjustment was low for all structures (<2 mm), although large local adjustment was observed for some structures. The median was systematically greater or equal to zero, indicating that the auto-contouring tends to under-segment the desired contour. CONCLUSION Auto-contouring performance assessment in routine clinical practice has identified systematic improvements required technically, but also highlighted the need for continued RTT training to ensure adherence to guidelines.
Collapse
Affiliation(s)
- Charlotte L. Brouwer
- University of Groningen, University Medical Center Groningen, Department of Radiation Oncology, Groningen, The Netherlands
| | | | | | | | - Roel J.H.M. Steenbakkers
- University of Groningen, University Medical Center Groningen, Department of Radiation Oncology, Groningen, The Netherlands
| | - Johannes A. Langendijk
- University of Groningen, University Medical Center Groningen, Department of Radiation Oncology, Groningen, The Netherlands
| | - Stefan Both
- University of Groningen, University Medical Center Groningen, Department of Radiation Oncology, Groningen, The Netherlands
| | | |
Collapse
|
41
|
Chaudhury A. Multilevel Optimization for Registration of Deformable Point Clouds. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2020; PP:8735-8746. [PMID: 32870792 DOI: 10.1109/tip.2020.3019649] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Handling deformation is one of the biggest challenges associated with point cloud registration. When deformation happens due to the motion of an animated object which actively changes its location and general shape, registration of two instances of the same object turns out to be a challenging task. The focus of this work is to address the problem by leveraging the complementary attributes of local and global geometric structures of the point clouds. We define an energy function which consists of local and global terms, as well as a semi-local term to model the intermediate level geometry of the point cloud. The local energy estimates the transformation parameters at the lowest level by assuming a reduced deformation model. The parameters are estimated in a closed form solution, which are then used to assign the initial probability of a stochastic model working at the intermediate level. The global energy term estimates the overall transformation parameters by minimizing a nonlinear least square function via Gauss-Newton optimization framework. The total energy is optimized in a block coordinate descent fashion, updating one term at a time while keeping others constant. Experiments on three publicly available datasets show that the method performs significantly better than several state-of-the-art algorithms in registering pairwise point cloud data.
Collapse
|
42
|
Du G, Wang K, Lian S, Zhao K. Vision-based robotic grasping from object localization, object pose estimation to grasp estimation for parallel grippers: a review. Artif Intell Rev 2020. [DOI: 10.1007/s10462-020-09888-5] [Citation(s) in RCA: 71] [Impact Index Per Article: 17.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/19/2022]
|
43
|
Efficient Similarity Point Set Registration by Transformation Decomposition. SENSORS 2020; 20:s20154103. [PMID: 32717938 PMCID: PMC7436047 DOI: 10.3390/s20154103] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/24/2020] [Revised: 07/20/2020] [Accepted: 07/22/2020] [Indexed: 11/25/2022]
Abstract
Point set registration is one of the basic problems in computer vision. When the overlap ratio between point sets is small or the relative transformation is large, local methods cannot guarantee the accuracy. However, the time complexity of the branch and bound (BnB) optimization used in most existing global methods is exponential in the dimensionality of parameter space. Therefore, seven-Degrees of Freedom (7-DoF) similarity transformation is a big challenge for BnB. In this paper, a novel rotation and scale invariant feature is introduced to decouple the optimization of translation, rotation, and scale in similarity point set registration, so that BnB optimization can be done in two lower dimensional spaces. With the transformation decomposition, the translation is first estimated and then the rotation is optimized by maximizing a robust objective function defined on consensus set. Finally, the scale is estimated according to the potential correspondences in the obtained consensus set. Experiments on synthetic data and clinical data show that our method is approximately two orders of magnitude faster than the state-of-the-art global method and more accurate than a typical local method. When the outlier ratio with respect to the inliers is up to 1.0, our method still achieves accurate registration.
Collapse
|
44
|
Khazari AE, Que Y, Sung TL, Lee HJ. Deep Global Features for Point Cloud Alignment. SENSORS (BASEL, SWITZERLAND) 2020; 20:s20144032. [PMID: 32698504 PMCID: PMC7411762 DOI: 10.3390/s20144032] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/15/2020] [Revised: 07/09/2020] [Accepted: 07/18/2020] [Indexed: 05/31/2023]
Abstract
Point cloud registration is a key problem in computer vision applications and involves finding a rigid transform from a point cloud into another such that they align together. The iterative closest point (ICP) method is a simple and effective solution that converges to a local optimum. However, despite the fact that point cloud registration or alignment is addressed in learning-based methods, such as PointNetLK, they do not offer good generalizability for point clouds. In this stud, we proposed a learning-based approach that addressed existing problems, such as finding local optima for ICP and achieving minimum generalizability. The proposed model consisted of three main parts: an encoding network, an auxiliary module that weighed the contribution of each input point cloud, and feature alignment to achieve the final transform. The proposed architecture offered greater generalization among the categories. Experiments were performed on ModelNet40 with different configurations and the results indicated that the proposed approach significantly outperformed the state-of-the-art point cloud alignment methods.
Collapse
Affiliation(s)
- Ahmed El Khazari
- Division of Computer Science and Engineering, Jeonbuk National University, Jeonju 54896, Korea; (A.E.K.); (Y.Q.); (T.L.S.)
| | - Yue Que
- Division of Computer Science and Engineering, Jeonbuk National University, Jeonju 54896, Korea; (A.E.K.); (Y.Q.); (T.L.S.)
| | - Thai Leang Sung
- Division of Computer Science and Engineering, Jeonbuk National University, Jeonju 54896, Korea; (A.E.K.); (Y.Q.); (T.L.S.)
| | - Hyo Jong Lee
- Division of Computer Science and Engineering, Jeonbuk National University, Jeonju 54896, Korea; (A.E.K.); (Y.Q.); (T.L.S.)
- Center for Advanced Image and Information Technology, Jeonbuk National University, Jeonju 54896, Korea
| |
Collapse
|
45
|
Zhang X, Wang T, Zhang X, Zhang Y, Wang J. Assessment and application of the coherent point drift algorithm to augmented reality surgical navigation for laparoscopic partial nephrectomy. Int J Comput Assist Radiol Surg 2020; 15:989-999. [PMID: 32361857 DOI: 10.1007/s11548-020-02163-6] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2019] [Accepted: 04/06/2020] [Indexed: 12/29/2022]
Abstract
PURPOSE The surface-based registration approach to laparoscopic augmented reality (AR) has clear advantages. Nonrigid point-set registration paves the way for surface-based registration. Among current non-rigid point set registration methods, the coherent point drift (CPD) algorithm is rarely used because of two challenges: (1) volumetric deformation is difficult to predict, and (2) registration from intraoperative visible tissue surface to whole anatomical preoperative model is a "part-to-whole" registration that CPD cannot be applied directly to. We preliminarily applied CPD on surgical navigation for laparoscopic partial nephrectomy (LPN). However, it introduces normalization errors and lacks navigation robustness. This paper presents important advances for more effectively applying CPD to LPN surgical navigation while attempting to quantitatively evaluate the accuracy of CPD-based surgical navigation. METHODS First, an optimized volumetric deformation (Op-VD) algorithm is proposed to achieve accurate prediction of volume deformation. Then, a projection-based partial selection method is presented to conveniently and robustly apply the CPD to LPN surgical navigation. Finally, kidneys with different deformations in vitro, phantom and in vivo experiments are performed to evaluate the accuracy and effectiveness of our approach. RESULTS The average root-mean-square error of volume deformation was refined to 0.84 mm. The mean target registration error (TRE) of the surface and inside markers in the in vitro experiments decreased to 1.51 mm and 1.29 mm, respectively. The robustness and precision of CPD-based navigation were validated in phantom and in vivo experiments, and the mean navigation TRE of the phantom experiments was found to be [Formula: see text] mm. CONCLUSION Accurate volumetric deformation and robust navigation results can be achieved in AR navigation of LPN by using surface-based registration with CPD. Evaluation results demonstrate the effectiveness of our proposed methods while showing the clinical application potential of CPD. This work has important guiding significance for the application of the CPD in laparoscopic AR.
Collapse
Affiliation(s)
- Xiaohui Zhang
- School of Mechanical Engineering and Automation, Beihang University, Beijing, China
| | - Tianmiao Wang
- School of Mechanical Engineering and Automation, Beihang University, Beijing, China
- Beijing Advanced Innovation Center for Biomedical Engineering, Beihang University, Beijing, China
| | - Xuebin Zhang
- Department of Urology, Peking Union Medical College Hospital, Peking Union Medical College and Chinese Academy of Medical Sciences, Beijing, China
| | - Yinghao Zhang
- School of Mechanical Engineering and Automation, Beihang University, Beijing, China
| | - Junchen Wang
- School of Mechanical Engineering and Automation, Beihang University, Beijing, China.
- Beijing Advanced Innovation Center for Biomedical Engineering, Beihang University, Beijing, China.
| |
Collapse
|
46
|
A robust and semi-automatic quantitative measurement of patellofemoral instability based on four dimensional computed tomography. Med Eng Phys 2020; 78:29-38. [PMID: 32115353 DOI: 10.1016/j.medengphy.2020.01.012] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2018] [Revised: 01/02/2020] [Accepted: 01/23/2020] [Indexed: 11/21/2022]
Abstract
Patellofemoral instability is a motion related disease, featured as the patella dislocating from the trochlear groove. Four dimensional computed tomography (4DCT) enables full assessment of the patellofemoral movement. Nevertheless, the quantitative measurements of patellofemoral instability are still under research and currently of limited practical use. The aim of this study is to develop a robust and semi-automatic workflow to quantitatively describe the patellofemoral movement in a patient group of eight suffering from patellofemoral instability. The initial results show agreement with manual observations of the tibial tubercle - trochlear groove (TT-TG) distance in routine practice, and the possibility to evaluate both TT-TG distance and patellar centre - trochlear groove (PC-TG) distance dynamically during active flexion-extension-flexion movement of the knee.
Collapse
|
47
|
Park BJ, Hunt SJ, Martin C, Nadolski GJ, Wood BJ, Gade TP. Augmented and Mixed Reality: Technologies for Enhancing the Future of IR. J Vasc Interv Radiol 2020; 31:1074-1082. [PMID: 32061520 DOI: 10.1016/j.jvir.2019.09.020] [Citation(s) in RCA: 52] [Impact Index Per Article: 13.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2019] [Revised: 08/01/2019] [Accepted: 09/20/2019] [Indexed: 10/25/2022] Open
Abstract
Augmented and mixed reality are emerging interactive and display technologies. These technologies are able to merge virtual objects, in either 2 or 3 dimensions, with the real world. Image guidance is the cornerstone of interventional radiology. With augmented or mixed reality, medical imaging can be more readily accessible or displayed in actual 3-dimensional space during procedures to enhance guidance, at times when this information is most needed. In this review, the current state of these technologies is addressed followed by a fundamental overview of their inner workings and challenges with 3-dimensional visualization. Finally, current and potential future applications in interventional radiology are highlighted.
Collapse
Affiliation(s)
- Brian J Park
- Department of Interventional Radiology, Hospital of the University of Pennsylvania, 3400 Spruce Street, Philadelphia, PA 19104.
| | - Stephen J Hunt
- Department of Interventional Radiology, Hospital of the University of Pennsylvania, 3400 Spruce Street, Philadelphia, PA 19104
| | - Charles Martin
- Department of Interventional Radiology, Cleveland Clinic, Cleveland, Ohio
| | - Gregory J Nadolski
- Department of Interventional Radiology, Hospital of the University of Pennsylvania, 3400 Spruce Street, Philadelphia, PA 19104
| | - Bradford J Wood
- Interventional Radiology, National Institutes of Health, Bethesda, Maryland
| | - Terence P Gade
- Department of Interventional Radiology, Hospital of the University of Pennsylvania, 3400 Spruce Street, Philadelphia, PA 19104
| |
Collapse
|
48
|
Moyou M, Rangarajan A, Corring J, Peter AM. A Grassmannian Graph Approach to Affine Invariant Feature Matching. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2019; 29:3374-3387. [PMID: 31880551 DOI: 10.1109/tip.2019.2959722] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
In this work, we present a novel, theoretical approach to address one of the longstanding problems in computer vision: 2D and 3D affine invariant feature matching. Our proposed Grassmannian Graph (GrassGraph) framework employs a two stage procedure that is capable of robustly recovering correspondences between two unorganized, affinely related feature (point) sets. In the ideal case, the first stage maps the feature sets to an affine invariant Grassmannian representation, where the features are mapped into the same subspace. It turns out that coordinate representations extracted from the Grassmannian differ by an arbitrary orthonormal matrix. In the second stage, by approximating the Laplace-Beltrami operator (LBO) on these coordinates, this extra orthonormal factor is nullified, providing true affine invariant coordinates which we then utilize to recover correspondences via simple mutual nearest neighbor relations. Our validation benchmarks use large number of experimental trials performed on 2D and 3D datasets. Experimental results show that the proposed GrassGraph method successfully recovers large affine transformations.
Collapse
|
49
|
Yang J, Quan S, Wang P, Zhang Y. Evaluating Local Geometric Feature Representations for 3D Rigid Data Matching. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2019; 29:2522-2535. [PMID: 31869788 DOI: 10.1109/tip.2019.2959236] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Local geometric descriptors act as an essential component for 3D rigid data matching. A rotational invariant local geometric descriptor usually consists of two components: local reference frame (LRF) and feature representation. However, existing evaluation efforts have mainly been paid on the LRF or the overall descriptor and the quantitative comparison of feature representations remains unexplored. This paper fills the gap by comprehensively evaluating nine state-of-the-art local geometric feature representations. In particular, our evaluation assesses feature representations based on ground-truth LRFs such that the ranking of tested methods is more convincing as compared with existing studies. The experiments are deployed on six standard datasets with various application scenarios (shape retrieval, point cloud registration, and object recognition) and data modalities (LiDAR, Kinect, and Space Time) as well as perturbations including Gaussian noise, shot noise, data decimation, clutter, occlusion, and limited overlap. The evaluated terms cover the major concerns for a feature representation, e.g., distinctiveness, robustness, compactness, and efficiency. The outcomes present interesting findings that may shed new light on this community and provide complementary perspectives to existing evaluations on the topic of local geometric feature description. A summary of evaluated methods regarding their peculiarities is finally presented to guide real-world applications and new descriptor crafting.
Collapse
|
50
|
|