26
|
Fuse S, Otake Y, Mifune Y, Tanaka H. A Facile Preparation of α-Aryl Carboxylic Acid via One-Flow Arndt–Eistert Synthesis. Aust J Chem 2015. [DOI: 10.1071/ch15342] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
Abstract
An efficient, one-flow Arndt–Eistert synthesis was demonstrated. A sequence of acid chloride formation–nucleophilic acyl substitution–Wolff rearrangement–nucleophilic addition was performed in a microflow system without isolating any intermediates, which included a potentially explosive compound. The microflow system was made from simple, inexpensive, and readily available instruments and tubes. α-Aryl esters 2a and 2b were prepared in yields of 33 and 23 % (three steps) respectively.
Collapse
|
27
|
Uneri A, Wang AS, Otake Y, Kleinszig G, Vogt S, Khanna AJ, Gallia GL, Gokaslan ZL, Siewerdsen JH. Evaluation of low-dose limits in 3D-2D rigid registration for surgical guidance. Phys Med Biol 2014; 59:5329-45. [PMID: 25146673 DOI: 10.1088/0031-9155/59/18/5329] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022]
|
28
|
Reaungamornrat S, Wang AS, Uneri A, Otake Y, Khanna AJ, Siewerdsen JH. Deformable image registration with local rigidity constraints for cone-beam CT-guided spine surgery. Phys Med Biol 2014; 59:3761-87. [PMID: 24937093 DOI: 10.1088/0031-9155/59/14/3761] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/27/2023]
Abstract
Image-guided spine surgery (IGSS) is associated with reduced co-morbidity and improved surgical outcome. However, precise localization of target anatomy and adjacent nerves and vessels relative to planning information (e.g., device trajectories) can be challenged by anatomical deformation. Rigid registration alone fails to account for deformation associated with changes in spine curvature, and conventional deformable registration fails to account for rigidity of the vertebrae, causing unrealistic distortions in the registered image that can confound high-precision surgery. We developed and evaluated a deformable registration method capable of preserving rigidity of bones while resolving the deformation of surrounding soft tissue. The method aligns preoperative CT to intraoperative cone-beam CT (CBCT) using free-form deformation (FFD) with constraints on rigid body motion imposed according to a simple intensity threshold of bone intensities. The constraints enforced three properties of a rigid transformation-namely, constraints on affinity (AC), orthogonality (OC), and properness (PC). The method also incorporated an injectivity constraint (IC) to preserve topology. Physical experiments involving phantoms, an ovine spine, and a human cadaver as well as digital simulations were performed to evaluate the sensitivity to registration parameters, preservation of rigid body morphology, and overall registration accuracy of constrained FFD in comparison to conventional unconstrained FFD (uFFD) and Demons registration. FFD with orthogonality and injectivity constraints (denoted FFD+OC+IC) demonstrated improved performance compared to uFFD and Demons. Affinity and properness constraints offered little or no additional improvement. The FFD+OC+IC method preserved rigid body morphology at near-ideal values of zero dilatation (D = 0.05, compared to 0.39 and 0.56 for uFFD and Demons, respectively) and shear (S = 0.08, compared to 0.36 and 0.44 for uFFD and Demons, respectively). Target registration error (TRE) was similarly improved for FFD+OC+IC (0.7 mm), compared to 1.4 and 1.8 mm for uFFD and Demons. Results were validated in human cadaver studies using CT and CBCT images, with FFD+OC+IC providing excellent preservation of rigid morphology and equivalent or improved TRE. The approach therefore overcomes distortions intrinsic to uFFD and could better facilitate high-precision IGSS.
Collapse
|
29
|
Wang A, Stayman J, Otake Y, Gallia G, Siewerdsen J. WE-G-18A-01: JUNIOR INVESTIGATOR WINNER - Low-Dose C-Arm Cone-Beam CT with Model-Based Image Reconstruction for High-Quality Guidance of Neurosurgical Intervention. Med Phys 2014. [DOI: 10.1118/1.4889512] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022] Open
|
30
|
Locatelli F, Choukroun G, Fliser D, Moecks J, Wiggenhauser A, Gupta A, Swinkels DW, Lin V, Guss C, Pratt R, Carrilho P, Martins AR, Alves M, Mateus A, Gusmao L, Parreira L, Assuncao J, Rodrigues I, Stamopoulos D, Mpakirtzi N, Afentakis N, Grapsa E, Zitt E, Sturm G, Kronenberg F, Neyer U, Knoll F, Lhotta K, Weiss G, Robinson BM, Larkina M, Bieber B, Kleophas W, Li Y, Locatelli F, McCullough K, Nolen JG, Port FK, Pisoni RL, Kalicki RM, Uehlinger DE, Ogawa C, Kanda F, Tomosugi N, Maeda T, Kuji T, Fujikawa T, Shino M, Shibata K, Kaneda T, Nishihara M, Satta H, Kawata SI, Koguchi N, Tamura K, Hirawa N, Toya Y, Umemura S, Chanliau J, Martin H, Stamatelou K, Gonzalez-Tabares L, Manamley N, Farouk M, Addison J, Donck J, Schneider A, Gutjahr-Lengsfeld L, Ritz E, Scharnagl H, Gelbrich G, Pilz S, Macdougall IC, Wanner C, Drechsler C, Kuntsevich V, Charen E, Kobena D, Sheth N, Siktel H, Levin NW, Winchester JF, Kotanko P, Kaysen G, Kuragano T, Kida A, Yahiro M, Nanami M, Nagasawa Y, Hasuike Y, Nakanishi T, Stamopoulos D, Mpakirtzi N, Dimitratou V, Griveas I, Lianos E, Grapsa E, Sasaki Y, Yamazaki S, Fujita K, Kurasawa M, Yorozu K, Shimonaka Y, Suzuki N, Yamamoto M, Zwiech R, Szczepa ska J, Bruzda-Zwiech A, Rao A, Gilg J, Caskey F, Kirkpantur A, Balci MM, Turkvatan A, Afsar B, Alkis M, Mandiroglu F, Kim YO, Yoon SA, Kim YS, Choi SJ, Min JW, Cheong MA, Hasuike Y, Kida A, Oue M, Yamamoto K, Kimura T, Fukao W, Yahiro M, Kaibe S, Nanami M, Nakanishi T, Djuric PS, Ikonomovski J, Tosic J, Jankovic A, Majster Z, Stankovic Popovic V, Dimkovic N, Aicardi Spalloni V, Del Vecchio L, Longhi S, Violo L, La Milia V, Pontoriero G, Locatelli F, Shino M, Kuji T, Fujikawa T, Toya Y, Umemura S, Macdougall I, Rumjon A, Mangahis E, Goldstein L, Ryzlewicz T, Becker F, Kilgallon W, Fukasawa M, Otake Y, Yamagishi T, Kamiyama M, Kobayashi H, Takeda M, Toida T, Sato Y, Fujimoto S. DIALYSIS ANAEMIA. Nephrol Dial Transplant 2014. [DOI: 10.1093/ndt/gfu176] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
|
31
|
Wang AS, Stayman JW, Otake Y, Khanna AJ, Gallia GL, Siewerdsen JH. Patient-Specific Minimum-Dose Imaging Protocols for Statistical Image Reconstruction in C-arm Cone-Beam CT Using Correlated Noise Injection. ACTA ACUST UNITED AC 2014; 9033. [PMID: 34211241 DOI: 10.1117/12.2043083] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/14/2022]
Abstract
Purpose A new method for accurately portraying the impact of low-dose imaging techniques in C-arm cone-beam CT (CBCT) is presented and validated, allowing identification of minimum-dose protocols suitable to a given imaging task on a patient-specific basis in scenarios that require repeat intraoperative scans. Method To accurately simulate lower-dose techniques and account for object-dependent noise levels (x-ray quantum noise and detector electronics noise) and correlations (detector blur), noise of the proper magnitude and correlation was injected into the projections from an initial CBCT acquired at the beginning of a procedure. The resulting noisy projections were then reconstructed to yield low-dose preview (LDP) images that accurately depict the image quality at any level of reduced dose in both filtered backprojection and statistical image reconstruction. Validation studies were conducted on a mobile C-arm, with the noise injection method applied to images of an anthropomorphic head phantom and cadaveric torso across a range of lower-dose techniques. Results Comparison of preview and real CBCT images across a full range of techniques demonstrated accurate noise magnitude (within ~5%) and correlation (matching noise-power spectrum, NPS). Other image quality characteristics (e.g., spatial resolution, contrast, and artifacts associated with beam hardening and scatter) were also realistically presented at all levels of dose and across reconstruction methods, including statistical reconstruction. Conclusion Generating low-dose preview images for a broad range of protocols gives a useful method to select minimum-dose techniques that accounts for complex factors of imaging task, patient-specific anatomy, and observer preference. The ability to accurately simulate the influence of low-dose acquisition in statistical reconstruction provides an especially valuable means of identifying low-dose limits in a manner that does not rely on a model for the nonlinear reconstruction process or a model of observer performance.
Collapse
|
32
|
Uneri A, Otake Y, Wang AS, Kleinszig G, Vogt S, Khanna AJ, Siewerdsen JH. 3D-2D registration for surgical guidance: effect of projection view angles on registration accuracy. Phys Med Biol 2013; 59:271-87. [PMID: 24351769 DOI: 10.1088/0031-9155/59/2/271] [Citation(s) in RCA: 35] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/22/2023]
Abstract
An algorithm for intensity-based 3D-2D registration of CT and x-ray projections is evaluated, specifically using single- or dual-projection views to provide 3D localization. The registration framework employs the gradient information similarity metric and covariance matrix adaptation evolution strategy to solve for the patient pose in six degrees of freedom. Registration performance was evaluated in an anthropomorphic phantom and cadaver, using C-arm projection views acquired at angular separation, Δθ, ranging from ∼0°-180° at variable C-arm magnification. Registration accuracy was assessed in terms of 2D projection distance error and 3D target registration error (TRE) and compared to that of an electromagnetic (EM) tracker. The results indicate that angular separation as small as Δθ ∼10°-20° achieved TRE <2 mm with 95% confidence, comparable or superior to that of the EM tracker. The method allows direct registration of preoperative CT and planning data to intraoperative fluoroscopy, providing 3D localization free from conventional limitations associated with external fiducial markers, stereotactic frames, trackers and manual registration.
Collapse
|
33
|
Togashi T, Takahashi E, Midorikawa K, Aoyama M, Yamakawa K, Sato T, Iwasaki A, Owada S, Yamanouchi K, Hara T, Matsubara S, Ohshima T, Otake Y, Tamasaku K, Tanaka H, Tanaka T, Tomizawa H, Watanabe T, Yabashi M, Ishikawa T. Extreme ultraviolet free electron laser seeded by high-order harmonic. Radiat Phys Chem Oxf Engl 1993 2013. [DOI: 10.1016/j.radphyschem.2013.06.013] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
34
|
Armiger RS, Otake Y, Iwaskiw AS, Wickwire AC, Ott KA, Voo LM, Armand M, Merkle AC. Biomechanical Response of Blast Loading to the Head Using 2D-3D Cineradiographic Registration. ACTA ACUST UNITED AC 2013. [DOI: 10.1007/978-3-319-00777-9_18] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/20/2023]
|
35
|
Reaungamornrat S, Liu WP, Wang AS, Otake Y, Nithiananthan S, Uneri A, Schafer S, Tryggestad E, Richmon J, Sorger JM, Siewerdsen JH, Taylor RH. Deformable image registration for cone-beam CT guided transoral robotic base-of-tongue surgery. Phys Med Biol 2013; 58:4951-79. [PMID: 23807549 PMCID: PMC3990286 DOI: 10.1088/0031-9155/58/14/4951] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/02/2023]
Abstract
Transoral robotic surgery (TORS) offers a minimally invasive approach to resection of base-of-tongue tumors. However, precise localization of the surgical target and adjacent critical structures can be challenged by the highly deformed intraoperative setup. We propose a deformable registration method using intraoperative cone-beam computed tomography (CBCT) to accurately align preoperative CT or MR images with the intraoperative scene. The registration method combines a Gaussian mixture (GM) model followed by a variation of the Demons algorithm. First, following segmentation of the volume of interest (i.e. volume of the tongue extending to the hyoid), a GM model is applied to surface point clouds for rigid initialization (GM rigid) followed by nonrigid deformation (GM nonrigid). Second, the registration is refined using the Demons algorithm applied to distance map transforms of the (GM-registered) preoperative image and intraoperative CBCT. Performance was evaluated in repeat cadaver studies (25 image pairs) in terms of target registration error (TRE), entropy correlation coefficient (ECC) and normalized pointwise mutual information (NPMI). Retraction of the tongue in the TORS operative setup induced gross deformation >30 mm. The mean TRE following the GM rigid, GM nonrigid and Demons steps was 4.6, 2.1 and 1.7 mm, respectively. The respective ECC was 0.57, 0.70 and 0.73, and NPMI was 0.46, 0.57 and 0.60. Registration accuracy was best across the superior aspect of the tongue and in proximity to the hyoid (by virtue of GM registration of surface points on these structures). The Demons step refined registration primarily in deeper portions of the tongue further from the surface and hyoid bone. Since the method does not use image intensities directly, it is suitable to multi-modality registration of preoperative CT or MR with intraoperative CBCT. Extending the 3D image registration to the fusion of image and planning data in stereo-endoscopic video is anticipated to support safer, high-precision base-of-tongue robotic surgery.
Collapse
|
36
|
Stayman JW, Dang H, Otake Y, Zbijewski W, Noble J, Dawant B, Labadie R, Carey JP, Siewerdsen JH. Overcoming Nonlinear Partial Volume Effects in Known-Component Reconstruction of Cochlear Implants. PROCEEDINGS OF SPIE--THE INTERNATIONAL SOCIETY FOR OPTICAL ENGINEERING 2013; 8668:86681L. [PMID: 24949189 PMCID: PMC4060628 DOI: 10.1117/12.2007945] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/14/2022]
Abstract
Nonlinear partial volume (NLPV) effects can be significant for objects with large attenuation differences and fine detail structures near the spatial resolution limits of a tomographic system. This is particularly true for small metal devices like cochlear implants. While traditional model-based approaches might alleviate these artifacts through very fine sampling of the image volume and subsampling of rays to each detector element, such solutions can be extremely burdensome in terms of memory and computational requirements. The work presented in this paper leverages the model-based approach called "known-component reconstruction" (KCR) where prior knowledge of a surgical device is integrated into the estimation. In KCR, the parameterization of the object separates the volume into an unknown background anatomy and a known component with unknown registration. Thus, one can model projections of an implant at very high spatial resolution while limiting the spatial resolution of the anatomy - in effect, modeling NLPV effects where they are most significant. We present modifications of the KCR approach that can be used to largely eliminate NLPV artifacts, and demonstrate the efficacy of the modified technique (with improved image quality and accurate implant position estimates) for the cochlear implant imaging scenario.
Collapse
|
37
|
Dang H, Otake Y, Schafer S, Stayman JW, Kleinszig G, Siewerdsen JH. Robust methods for automatic image-to-world registration in cone-beam CT interventional guidance. Med Phys 2012; 39:6484-98. [PMID: 23039683 DOI: 10.1118/1.4754589] [Citation(s) in RCA: 24] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022] Open
Abstract
PURPOSE Real-time surgical navigation relies on accurate image-to-world registration to align the coordinate systems of the image and patient. Conventional manual registration can present a workflow bottleneck and is prone to manual error and intraoperator variability. This work reports alternative means of automatic image-to-world registration, each method involving an automatic registration marker (ARM) used in conjunction with C-arm cone-beam CT (CBCT). The first involves a Known-Model registration method in which the ARM is a predefined tool, and the second is a Free-Form method in which the ARM is freely configurable. METHODS Studies were performed using a prototype C-arm for CBCT and a surgical tracking system. A simple ARM was designed with markers comprising a tungsten sphere within infrared reflectors to permit detection of markers in both x-ray projections and by an infrared tracker. The Known-Model method exercised a predefined specification of the ARM in combination with 3D-2D registration to estimate the transformation that yields the optimal match between forward projection of the ARM and the measured projection images. The Free-Form method localizes markers individually in projection data by a robust Hough transform approach extended from previous work, backprojected to 3D image coordinates based on C-arm geometric calibration. Image-domain point sets were transformed to world coordinates by rigid-body point-based registration. The robustness and registration accuracy of each method was tested in comparison to manual registration across a range of body sites (head, thorax, and abdomen) of interest in CBCT-guided surgery, including cases with interventional tools in the radiographic scene. RESULTS The automatic methods exhibited similar target registration error (TRE) and were comparable or superior to manual registration for placement of the ARM within ∼200 mm of C-arm isocenter. Marker localization in projection data was robust across all anatomical sites, including challenging scenarios involving the presence of interventional tools. The reprojection error of marker localization was independent of the distance of the ARM from isocenter, and the overall TRE was dominated by the configuration of individual fiducials and distance from the target as predicted by theory. The median TRE increased with greater ARM-to-isocenter distance (e.g., for the Free-Form method, TRE increasing from 0.78 mm to 2.04 mm at distances of ∼75 mm and 370 mm, respectively). The median TRE within ∼200 mm distance was consistently lower than that of the manual method (TRE = 0.82 mm). Registration performance was independent of anatomical site (head, thorax, and abdomen). The Free-Form method demonstrated a statistically significant improvement (p = 0.0044) in reproducibility compared to manual registration (0.22 mm versus 0.30 mm, respectively). CONCLUSIONS Automatic image-to-world registration methods demonstrate the potential for improved accuracy, reproducibility, and workflow in CBCT-guided procedures. A Free-Form method was shown to exhibit robustness against anatomical site, with comparable or improved TRE compared to manual registration. It was also comparable or superior in performance to a Known-Model method in which the ARM configuration is specified as a predefined tool, thereby allowing configuration of fiducials on the fly or attachment to the patient.
Collapse
|
38
|
Otake Y, Schafer S, Stayman JW, Zbijewski W, Kleinszig G, Graumann R, Khanna AJ, Siewerdsen JH. Automatic localization of vertebral levels in x-ray fluoroscopy using 3D-2D registration: a tool to reduce wrong-site surgery. Phys Med Biol 2012; 57:5485-508. [PMID: 22864366 DOI: 10.1088/0031-9155/57/17/5485] [Citation(s) in RCA: 50] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
Abstract
Surgical targeting of the incorrect vertebral level (wrong-level surgery) is among the more common wrong-site surgical errors, attributed primarily to the lack of uniquely identifiable radiographic landmarks in the mid-thoracic spine. The conventional localization method involves manual counting of vertebral bodies under fluoroscopy, is prone to human error and carries additional time and dose. We propose an image registration and visualization system (referred to as LevelCheck), for decision support in spine surgery by automatically labeling vertebral levels in fluoroscopy using a GPU-accelerated, intensity-based 3D-2D (namely CT-to-fluoroscopy) registration. A gradient information (GI) similarity metric and a CMA-ES optimizer were chosen due to their robustness and inherent suitability for parallelization. Simulation studies involved ten patient CT datasets from which 50 000 simulated fluoroscopic images were generated from C-arm poses selected to approximate the C-arm operator and positioning variability. Physical experiments used an anthropomorphic chest phantom imaged under real fluoroscopy. The registration accuracy was evaluated as the mean projection distance (mPD) between the estimated and true center of vertebral levels. Trials were defined as successful if the estimated position was within the projection of the vertebral body (namely mPD <5 mm). Simulation studies showed a success rate of 99.998% (1 failure in 50 000 trials) and computation time of 4.7 s on a midrange GPU. Analysis of failure modes identified cases of false local optima in the search space arising from longitudinal periodicity in vertebral structures. Physical experiments demonstrated the robustness of the algorithm against quantum noise and x-ray scatter. The ability to automatically localize target anatomy in fluoroscopy in near-real-time could be valuable in reducing the occurrence of wrong-site surgery while helping to reduce radiation exposure. The method is applicable beyond the specific case of vertebral labeling, since any structure defined in pre-operative (or intra-operative) CT or cone-beam CT can be automatically registered to the fluoroscopic scene.
Collapse
|
39
|
Zbijewski W, Stayman J, Otake Y, Carrino J, Khanna A, Siewerdsen J. WE-G-217BCD-01: BEST IN PHYSICS (IMAGING) - High-Quality CT Imaging in the Presence of Surgical Instrumentation Using Spectral System Models and Knowledge of Implanted Devices. Med Phys 2012; 39:3972-3973. [DOI: 10.1118/1.4736211] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022] Open
|
40
|
Yoo J, Schafer S, Uneri A, Otake Y, Khanna AJ, Siewerdsen JH. An electromagnetic “Tracker-in-Table” configuration for X-ray fluoroscopy and cone-beam CT-guided surgery. Int J Comput Assist Radiol Surg 2012; 8:1-13. [PMID: 22585463 DOI: 10.1007/s11548-012-0744-z] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2011] [Accepted: 04/26/2012] [Indexed: 11/25/2022]
|
41
|
Reaungamornrat S, Otake Y, Uneri A, Schafer S, Mirota DJ, Nithiananthan S, Stayman JW, Kleinszig G, Khanna AJ, Taylor RH, Siewerdsen JH. An on-board surgical tracking and video augmentation system for C-arm image guidance. Int J Comput Assist Radiol Surg 2012; 7:647-65. [PMID: 22539008 DOI: 10.1007/s11548-012-0682-9] [Citation(s) in RCA: 23] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2011] [Accepted: 03/20/2012] [Indexed: 11/27/2022]
Abstract
PURPOSE Conventional tracker configurations for surgical navigation carry a variety of limitations, including limited geometric accuracy, line-of-sight obstruction, and mismatch of the view angle with the surgeon's-eye view. This paper presents the development and characterization of a novel tracker configuration (referred to as "Tracker-on-C") intended to address such limitations by incorporating the tracker directly on the gantry of a mobile C-arm for fluoroscopy and cone-beam CT (CBCT). METHODS A video-based tracker (MicronTracker, Claron Technology Inc., Toronto, ON, Canada) was mounted on the gantry of a prototype mobile isocentric C-arm next to the flat-panel detector. To maintain registration within a dynamically moving reference frame (due to rotation of the C-arm), a reference marker consisting of 6 faces (referred to as a "hex-face marker") was developed to give visibility across the full range of C-arm rotation. Three primary functionalities were investigated: surgical tracking, generation of digitally reconstructed radiographs (DRRs) from the perspective of a tracked tool or the current C-arm angle, and augmentation of the tracker video scene with image, DRR, and planning data. Target registration error (TRE) was measured in comparison with the same tracker implemented in a conventional in-room configuration. Graphics processing unit (GPU)-accelerated DRRs were generated in real time as an assistant to C-arm positioning (i.e., positioning the C-arm such that target anatomy is in the field-of-view (FOV)), radiographic search (i.e., a virtual X-ray projection preview of target anatomy without X-ray exposure), and localization (i.e., visualizing the location of the surgical target or planning data). Video augmentation included superimposing tracker data, the X-ray FOV, DRRs, planning data, preoperative images, and/or intraoperative CBCT onto the video scene. Geometric accuracy was quantitatively evaluated in each case, and qualitative assessment of clinical feasibility was analyzed by an experienced and fellowship-trained orthopedic spine surgeon within a clinically realistic surgical setup of the Tracker-on-C. RESULTS The Tracker-on-C configuration demonstrated improved TRE (0.87 ± 0.25) mm in comparison with a conventional in-room tracker setup (1.92 ± 0.71) mm (p < 0.0001) attributed primarily to improved depth resolution of the stereoscopic camera placed closer to the surgical field. The hex-face reference marker maintained registration across the 180° C-arm orbit (TRE = 0.70 ± 0.32 mm). DRRs generated from the perspective of the C-arm X-ray detector demonstrated sub- mm accuracy (0.37 ± 0.20 mm) in correspondence with the real X-ray image. Planning data and DRRs overlaid on the video scene exhibited accuracy of (0.59 ± 0.38) pixels and (0.66 ± 0.36) pixels, respectively. Preclinical assessment suggested potential utility of the Tracker-on-C in a spectrum of interventions, including improved line of sight, an assistant to C-arm positioning, and faster target localization, while reducing X-ray exposure time. CONCLUSIONS The proposed tracker configuration demonstrated sub- mm TRE from the dynamic reference frame of a rotational C-arm through the use of the multi-face reference marker. Real-time DRRs and video augmentation from a natural perspective over the operating table assisted C-arm setup, simplified radiographic search and localization, and reduced fluoroscopy time. Incorporation of the proposed tracker configuration with C-arm CBCT guidance has the potential to simplify intraoperative registration, improve geometric accuracy, enhance visualization, and reduce radiation exposure.
Collapse
|
42
|
Stayman JW, Otake Y, Schafer S, Khanna AJ, Prince JL, Siewerdsen JH. Model-based Reconstruction of Objects with Inexactly Known Components. PROCEEDINGS OF SPIE--THE INTERNATIONAL SOCIETY FOR OPTICAL ENGINEERING 2012. [PMID: 26203201 DOI: 10.1117/12.911202] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/14/2022]
Abstract
Because tomographic reconstructions are ill-conditioned, algorithms that incorporate additional knowledge about the imaging volume generally have improved image quality. This is particularly true when measurements are noisy or have missing data. This paper presents a general reconstruction framework for including attenuation contributions from objects known to be in the field-of-view. Components such as surgical devices and tools may be modeled explicitly as part of the attenuating volume but are inexactly known with respect to their locations poses, and possible deformations. The proposed reconstruction framework, referred to as Known-Component Reconstruction (KCR), is based on this novel parameterization of the object, a likelihood-based objective function, and alternating optimizations between registration and image parameters to jointly estimate the both the underlying attenuation and unknown registrations. A deformable KCR (dKCR) approach is introduced that adopts a control point-based warping operator to accommodate shape mismatches between the component model and the physical component, thereby allowing for a more general class of inexactly known components. The KCR and dKCR approaches are applied to low-dose cone-beam CT data with spine fixation hardware present in the imaging volume. Such data is particularly challenging due to photon starvation effects in projection data behind the metallic components. The proposed algorithms are compared with traditional filtered-backprojection and penalized-likelihood reconstructions and found to provide substantially improved image quality. Whereas traditional approaches exhibit significant artifacts that complicate detection of breaches or fractures near metal, the KCR framework tends to provide good visualization of anatomy right up to the boundary of surgical devices.
Collapse
|
43
|
Lee J, Stayman JW, Otake Y, Schafer S, Zbijewski W, Khanna AJ, Prince JL, Siewerdsen JH. Incorporation of Prior Knowledge for Region of Change Imaging from Sparse Scan Data in Image-Guided Surgery. PROCEEDINGS OF SPIE--THE INTERNATIONAL SOCIETY FOR OPTICAL ENGINEERING 2012; 8316:831603. [PMID: 26166930 PMCID: PMC4497550 DOI: 10.1117/12.910850] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
This paper proposes to utilize a patient-specific prior to augment intraoperative sparse-scan data to accurately reconstruct the aspects of the region that have changed by a surgical procedure in image-guided surgeries. When anatomical changes are introduced by a surgical procedure, only a sparse set of x-ray images are acquired, and the prior volume is registered to these data. Since all the information of the patient anatomy except for the surgical change is already known from the prior volume, we highlight only the change by creating difference images between the new scan and digitally reconstructed radiographs (DRR) computed from the registered prior volume. The region of change (RoC) is reconstructed from these sparse difference images by a penalized likelihood (PL) reconstruction method regularized by a compressed sensing penalty. When the surgical changes are local and relatively small, the RoC reconstruction involves only a small volume size and a small number of projections, allowing much faster computation and lower radiation dose than is needed to reconstruct the entire surgical volume. The reconstructed RoC merges with the prior volume to visualize an updated surgical field. We apply this novel approach to sacroplasty phantom data obtained from a cone-beam CT (CBCT) test bench and vertebroplasty data with a fresh cadaver acquired from a C-arm CBCT system with a flat-panel detector (FPD).
Collapse
|
44
|
Uneri A, Schafer S, Mirota DJ, Nithiananthan S, Otake Y, Taylor RH, Gallia GL, Khanna AJ, Lee S, Reh DD, Siewerdsen JH. TREK: an integrated system architecture for intraoperative cone-beam CT-guided surgery. Int J Comput Assist Radiol Surg 2011; 7:159-73. [PMID: 21744085 DOI: 10.1007/s11548-011-0636-7] [Citation(s) in RCA: 74] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2011] [Accepted: 06/10/2011] [Indexed: 10/18/2022]
Abstract
PURPOSE A system architecture has been developed for integration of intraoperative 3D imaging [viz., mobile C-arm cone-beam CT (CBCT)] with surgical navigation (e.g., trackers, endoscopy, and preoperative image and planning data). The goal of this paper is to describe the architecture and its handling of a broad variety of data sources in modular tool development for streamlined use of CBCT guidance in application-specific surgical scenarios. METHODS The architecture builds on two proven open-source software packages, namely the cisst package (Johns Hopkins University, Baltimore, MD) and 3D Slicer (Brigham and Women's Hospital, Boston, MA), and combines data sources common to image-guided procedures with intraoperative 3D imaging. Integration at the software component level is achieved through language bindings to a scripting language (Python) and an object-oriented approach to abstract and simplify the use of devices with varying characteristics. The platform aims to minimize offline data processing and to expose quantitative tools that analyze and communicate factors of geometric precision online. Modular tools are defined to accomplish specific surgical tasks, demonstrated in three clinical scenarios (temporal bone, skull base, and spine surgery) that involve a progressively increased level of complexity in toolset requirements. RESULTS The resulting architecture (referred to as "TREK") hosts a collection of modules developed according to application-specific surgical tasks, emphasizing streamlined integration with intraoperative CBCT. These include multi-modality image display; 3D-3D rigid and deformable registration to bring preoperative image and planning data to the most up-to-date CBCT; 3D-2D registration of planning and image data to real-time fluoroscopy; infrared, electromagnetic, and video-based trackers used individually or in hybrid arrangements; augmented overlay of image and planning data in endoscopic or in-room video; and real-time "virtual fluoroscopy" computed from GPU-accelerated digitally reconstructed radiographs (DRRs). Application in three preclinical scenarios (temporal bone, skull base, and spine surgery) demonstrates the utility of the modular, task-specific approach in progressively complex tasks. CONCLUSIONS The design and development of a system architecture for image-guided surgery has been reported, demonstrating enhanced utilization of intraoperative CBCT in surgical applications with vastly different requirements. The system integrates C-arm CBCT with a broad variety of data sources in a modular fashion that streamlines the interface to application-specific tools, accommodates distinct workflow scenarios, and accelerates testing and translation of novel toolsets to clinical use. The modular architecture was shown to adapt to and satisfy the requirements of distinct surgical scenarios from a common code-base, leveraging software components arising from over a decade of effort within the imaging and computer-assisted interventions community.
Collapse
|
45
|
Stayman J, Otake Y, Uneri A, Prince J, Siewerdsen J. WE-A-301-04: Model-Based Known Component Reconstruction for Computed Tomography. Med Phys 2011. [DOI: 10.1118/1.3613288] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022] Open
|
46
|
Schafer S, Otake Y, Uneri A, Mirota D, Nithiananthan S, Stayman J, Zbijewski W, Schmidgunst C, Kleinszig G, Siewerdsen J. TH-E-110-09: Grids Revisited: The Effect of Antiscatter Grids on Image Quality and Dose in Mobile C-Arm Cone-Beam CT for Image-Guided Surgery. Med Phys 2011. [DOI: 10.1118/1.3613593] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022] Open
|
47
|
Nithiananthan S, Schafer S, Mirota DJ, Uneri A, Otake Y, Stayman JW, Siewerdsen JH. SU-E-J-47: Deformable Image Registration in the Presence of Excised Tissue: A Modified Demons Algorithm for Cone-Beam CT-Guided Surgery. Med Phys 2011. [DOI: 10.1118/1.3611815] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022] Open
|
48
|
Stayman J, Zbijewski W, Otake Y, Schafer S, Lee J, Prince J, Siewerdsen J. WE-A-301-07: Using Prior Images with Registration in Penalized Likelihood Estimation for CT with Sparse Data. Med Phys 2011. [DOI: 10.1118/1.3613291] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022] Open
|
49
|
Nithiananthan S, Mirota D, Uneri A, Schafer S, Otake Y, Stayman JW, Siewerdsen JH. Incorporating Tissue Excision in Deformable Image Registration: A Modified Demons Algorithm for Cone-Beam CT-Guided Surgery. PROCEEDINGS OF SPIE--THE INTERNATIONAL SOCIETY FOR OPTICAL ENGINEERING 2011; 7964:796404. [PMID: 37621997 PMCID: PMC10448501 DOI: 10.1117/12.878258] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/26/2023]
Abstract
The ability to perform fast, accurate, deformable registration with intraoperative images featuring surgical excisions was investigated for use in cone-beam CT (CBCT) guided head and neck surgery. Existing deformable registration methods generally fail to account for tissue excised between image acquisitions and typically simply "move" voxels within the images with no ability to account for tissue that is removed (or introduced) between scans. We have thus developed an approach in which an extra dimension is added during the registration process to act as a sink for voxels removed during the course of the procedure. A series of cadaveric images acquired using a prototype CBCT-capable C-arm were used to model tissue deformation and excision occurring during a surgical procedure, and the ability of deformable registration to correctly account for anatomical changes under these conditions was investigated. Using a previously developed version of the Demons deformable registration algorithm, we identify the difficulties that traditional registration algorithms encounter when faced with excised tissue and present a modified version of the algorithm better suited for use in intraoperative image-guided procedures. Studies were performed for different deformation and tissue excision tasks, and registration performance was quantified in terms of the ability to accurately account for tissue excision while avoiding spurious deformations arising around the excision.
Collapse
|
50
|
Stayman J, Zbijewski W, Otake Y, Prince J, Siewerdsen J. TH-D-201B-07: Predicting Noise and Resolution Properties in Tomosynthesis with Statistical Image Reconstruction. Med Phys 2010. [DOI: 10.1118/1.3469566] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022] Open
|