1
|
Zhu M, He B, Yu J, Yuan F, Liu J. HydraMarker: Efficient, Flexible, and Multifold Marker Field Generation. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2023; 45:5849-5861. [PMID: 36215370 DOI: 10.1109/tpami.2022.3212862] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
An n-order marker field is a special binary matrix whose n×n subregions are all distinct from each other in four orientations. It is commonly used to guide the composing process of position-sensing markers, which can be detected and identified in a camera image with very limited scope or severe visibility problems. Despite the advantages, position-sensing markers are rare and overlooked because generating marker fields is difficult. In this article, we broaden the definition of marker field, making it more powerful and flexible. Then, we propose bWFC (binary wave function collapse) and its high-speed version, fast-bWFC, to solve the generation problem. The methods are packaged into an open-sourced toolkit named HydraMarker, with which, users not only can generate marker fields on laptops within a short period of time, but also can highly customize them: preset values; fields and subregions in any shape; multifold local uniqueness. Comparative results indicate that the proposed method has superior efficiency, quality, and capability. It makes marker field generation accessible to common marker designers, opening up more possibilities for fiducial markers.
Collapse
|
2
|
Marchionna L, Pugliese G, Martini M, Angarano S, Salvetti F, Chiaberge M. Deep Instance Segmentation and Visual Servoing to Play Jenga with a Cost-Effective Robotic System. SENSORS (BASEL, SWITZERLAND) 2023; 23:752. [PMID: 36679543 PMCID: PMC9866192 DOI: 10.3390/s23020752] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/14/2022] [Revised: 01/04/2023] [Accepted: 01/06/2023] [Indexed: 06/17/2023]
Abstract
The game of Jenga is a benchmark used for developing innovative manipulation solutions for complex tasks. Indeed, it encourages the study of novel robotics methods to successfully extract blocks from a tower. A Jenga game involves many traits of complex industrial and surgical manipulation tasks, requiring a multi-step strategy, the combination of visual and tactile data, and the highly precise motion of a robotic arm to perform a single block extraction. In this work, we propose a novel, cost-effective architecture for playing Jenga with e.Do, a 6DOF anthropomorphic manipulator manufactured by Comau, a standard depth camera, and an inexpensive monodirectional force sensor. Our solution focuses on a visual-based control strategy to accurately align the end-effector with the desired block, enabling block extraction by pushing. To this aim, we trained an instance segmentation deep learning model on a synthetic custom dataset to segment each piece of the Jenga tower, allowing for visual tracking of the desired block's pose during the motion of the manipulator. We integrated the visual-based strategy with a 1D force sensor to detect whether the block could be safely removed by identifying a force threshold value. Our experimentation shows that our low-cost solution allows e.DO to precisely reach removable blocks and perform up to 14 consecutive extractions in a row.
Collapse
Affiliation(s)
- Luca Marchionna
- Department of Electronics and Telecommunications (DET), Politecnico di Torino, 10129 Torino, Italy
| | - Giulio Pugliese
- Department of Electronics and Telecommunications (DET), Politecnico di Torino, 10129 Torino, Italy
| | - Mauro Martini
- Department of Electronics and Telecommunications (DET), Politecnico di Torino, 10129 Torino, Italy
- PIC4SeR Interdepartmental Centre for Service Robotics, 10129 Torino, Italy
| | - Simone Angarano
- Department of Electronics and Telecommunications (DET), Politecnico di Torino, 10129 Torino, Italy
- PIC4SeR Interdepartmental Centre for Service Robotics, 10129 Torino, Italy
| | - Francesco Salvetti
- Department of Electronics and Telecommunications (DET), Politecnico di Torino, 10129 Torino, Italy
- PIC4SeR Interdepartmental Centre for Service Robotics, 10129 Torino, Italy
| | - Marcello Chiaberge
- Department of Electronics and Telecommunications (DET), Politecnico di Torino, 10129 Torino, Italy
- PIC4SeR Interdepartmental Centre for Service Robotics, 10129 Torino, Italy
| |
Collapse
|
3
|
Planning and visual-servoing for robotic manipulators in ROS. INTERNATIONAL JOURNAL OF INTELLIGENT ROBOTICS AND APPLICATIONS 2022. [DOI: 10.1007/s41315-022-00253-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/15/2022]
Abstract
AbstractThis article presents a probabilistic road map (PRM) and visual servo control (visual-servoing) based path planning strategy that allows a Motoman HP20D industrial robot to move from an initial positional to a random final position in the presence of fixed obstacles. The process begins with an application of the PRM algorithm to take the robot from an initial position to a point in space where it has a free line of sight to the target, to then apply visual servoing and end up, finally, at the desired position, where an image captured by a camera located at the robot’s end effector matches a reference image, located on the upper surface of a rectangular prismatic object. Algorithms and experiments were developed in simulation, specifically, the visual servo control that includes the dynamic model of the robot and the image sensor subject to realistic lighting were developed in robot operating system (ROS) environment.
Collapse
|
4
|
Dong L, Morel G. Robust trocar identification and its application in robotic minimally invasive surgery. Int J Med Robot 2022; 18:e2392. [PMID: 35368139 DOI: 10.1002/rcs.2392] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2022] [Revised: 02/15/2022] [Accepted: 03/11/2022] [Indexed: 11/09/2022]
Abstract
BACKGROUND In minimally invasive surgery (MIS), instruments pass through trocars which are installed through the incision points. This forms a fulcrum effect and imposes significant constraint. For robotic manipulative operations, the real-time trocar information is a prerequisite. Systems acquire this knowledge either with a prior registration procedure or through coordinated control of their joints. METHODS A robust and real-time trocar identification algorithm based on least square (LS) algorithm was proposed in the context of human-robot co-manipulation scenario. RESULTS Both in vitro and in vivo experiments were performed to verify the effectiveness of the proposed algorithm. The estimated trocar coordinates expressed in the robot base frame were further leveraged to implement an instrument gravity compensation function. CONCLUSIONS An LS based approach can be employed to robustly estimate the real-time trocar information so as to implement more practical robotic functions.
Collapse
Affiliation(s)
- Lin Dong
- Institut des Systèmes Intelligents et de Robotique, Sorbonne Université, CNRS, INSERM, Paris, France
| | - Guillaume Morel
- Institut des Systèmes Intelligents et de Robotique, Sorbonne Université, CNRS, INSERM, Paris, France
| |
Collapse
|
5
|
Ramesh A, Beniwal M, Uppar AM, V V, Rao M. Microsurgical Tool Detection and Characterization in Intra-operative Neurosurgical Videos. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2021; 2021:2676-2681. [PMID: 34891803 DOI: 10.1109/embc46164.2021.9630274] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Brain surgery is complex and has evolved as a separate surgical specialty. Surgical procedures on the brain are performed using dedicated micro-instruments which are designed specifically for the requirements of operating with finesse in a confined space. The usage of these microsurgical tools in an operating environment defines the surgical skill of a surgeon. Video recordings of micro-surgical procedures are a rich source of information to develop automated surgical assessment tools that can offer continuous feedback for surgeons to improve their skills, effectively increase the outcome of the surgery, and make a positive impact on their patients. This work presents a novel deep learning system based on the Yolov5 algorithm to automatically detect, localize and characterize microsurgical tools from recorded intra-operative neurosurgical videos. The tool detection achieves a high 93.2% mean average precision. The detected tools are then characterized by their on-off time, motion trajectory and usage time. Tool characterization from neurosurgical videos offers useful insight into the surgical methods employed by a surgeon and can aid in their improvement. Additionally, a new dataset of annotated neurosurgical videos is used to develop the robust model and is made available for the research community.Clinical relevance- Tool detection and characterization in neurosurgery has several online and offline applications including skill assessment and outcome of the surgery. The development of automated tool characterization systems for intra-operative neurosurgery is expected to not only improve the surgical skills of the surgeon, but also leverage in training the neurosurgical workforce. Additionally, dedicated neurosurgical video based datasets will, in general, aid the research community to explore more automation in this field.
Collapse
|
6
|
Xue Y, Li Y, Liu S, Wang P, Qian X. Oriented Localization of Surgical Tools by Location Encoding. IEEE Trans Biomed Eng 2021; 69:1469-1480. [PMID: 34652994 DOI: 10.1109/tbme.2021.3120430] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Surgical tool localization is the foundation to a series of advanced surgical functions e.g. image guided surgical navigation. For precise scenarios like surgical tool localization, sophisticated tools and sensitive tissues can be quite close. This requires a higher localization accuracy than general object localization. And it is also meaningful to know the orientation of tools. To achieve these, this paper proposes a Compressive Sensing based Location Encoding scheme, which formulates the task of surgical tool localization in pixel space into a task of vector regression in encoding space. Furthermore with this scheme, the method is able to capture orientation of surgical tools rather than simply outputting horizontal bounding boxes. To prevent gradient vanishing, a novel back-propagation rule for sparse reconstruction is derived. The back-propagation rule is applicable to different implementations of sparse reconstruction and renders the entire network end-to-end trainable. Finally, the proposed approach gives more accurate bounding boxes as well as capturing the orientation of tools, and achieves state-of-the-art performance compared with 9 competitive both oriented and non-oriented localization methods (RRD, RefineDet, etc) on a mainstream surgical image dataset: m2cai16-tool-locations. A range of experiments support our claim that regression in CSLE space performs better than traditionally detecting bounding boxes in pixel space.
Collapse
|
7
|
Sandoval J, Laribi MA, Faure JP, Breque C, Richer JP, Zeghloul S. Towards an Autonomous Robot-Assistant for Laparoscopy Using Exteroceptive Sensors: Feasibility Study and Implementation. IEEE Robot Autom Lett 2021. [DOI: 10.1109/lra.2021.3094644] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
8
|
Ma G, Ross W, Codd PJ. StereoCNC: A Stereovision-guided Robotic Laser System. PROCEEDINGS OF THE ... IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS. IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS 2021; 2021:540-547. [PMID: 35950084 PMCID: PMC9358620 DOI: 10.1109/iros51168.2021.9636050] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
This paper proposes an End-to-End stereovision-guided laser surgery system that can conduct laser ablation on targets selected by human operators in the color image, referred as StereoCNC. Two digital cameras are integrated into a previously developed robotic laser system to add a color sensing modality and formulate the stereovision. A calibration method is implemented to register the coordinate frames between stereo cameras and the laser system, modelled as a 3D-to-3D least-squares problem. The calibration reprojection errors are used to characterize a 3D error field by Gaussian Process Regression (GPR). This error field can make predictions for new point cloud data to identify an optimal position with lower calibration errors. A stereovision-guided laser ablation pipeline is proposed to optimize the positioning of the surgical site within the error field, which is achieved with a Genetic Algorithm search; mechanical stages move the site to the low-error region. The pipeline is validated by the experiments on phantoms with color texture and various geometric shapes. The overall targeting accuracy of the system achieved an average RMSE of 0.13 ± 0.02 mm and maximum error of 0.34 ± 0.06 mm, as measured by pre- and post-laser ablation images. The results show potential applications of using the developed stereovision-guided robotic system for superficial laser surgery, including dermatologic applications or removal of exposed tumorous tissue in neurosurgery.
Collapse
Affiliation(s)
- Guangshen Ma
- Brain Tool Lab. Department of Mechanical Engineering, Duke University
| | - Weston Ross
- Brain Tool Lab. Department of Mechanical Engineering, Duke University
- Department of Neurosurgery, Duke University
| | - Patrick J Codd
- Brain Tool Lab. Department of Mechanical Engineering, Duke University
- Department of Neurosurgery, Duke University
| |
Collapse
|
9
|
Design and Evaluation of a Foot-Controlled Robotic System for Endoscopic Surgery. IEEE Robot Autom Lett 2021. [DOI: 10.1109/lra.2021.3062009] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
|
10
|
Hasan MK, Calvet L, Rabbani N, Bartoli A. Detection, segmentation, and 3D pose estimation of surgical tools using convolutional neural networks and algebraic geometry. Med Image Anal 2021; 70:101994. [PMID: 33611053 DOI: 10.1016/j.media.2021.101994] [Citation(s) in RCA: 26] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2020] [Revised: 01/27/2021] [Accepted: 02/01/2021] [Indexed: 02/06/2023]
Abstract
BACKGROUND AND OBJECTIVE Surgical tool detection, segmentation, and 3D pose estimation are crucial components in Computer-Assisted Laparoscopy (CAL). The existing frameworks have two main limitations. First, they do not integrate all three components. Integration is critical; for instance, one should not attempt computing pose if detection is negative. Second, they have highly specific requirements, such as the availability of a CAD model. We propose an integrated and generic framework whose sole requirement for the 3D pose is that the tool shaft is cylindrical. Our framework makes the most of deep learning and geometric 3D vision by combining a proposed Convolutional Neural Network (CNN) with algebraic geometry. We show two applications of our framework in CAL: tool-aware rendering in Augmented Reality (AR) and tool-based 3D measurement. METHODS We name our CNN as ART-Net (Augmented Reality Tool Network). It has a Single Input Multiple Output (SIMO) architecture with one encoder and multiple decoders to achieve detection, segmentation, and geometric primitive extraction. These primitives are the tool edge-lines, mid-line, and tip. They allow the tool's 3D pose to be estimated by a fast algebraic procedure. The framework only proceeds if a tool is detected. The accuracy of segmentation and geometric primitive extraction is boosted by a new Full resolution feature map Generator (FrG). We extensively evaluate the proposed framework with the EndoVis and new proposed datasets. We compare the segmentation results against several variants of the Fully Convolutional Network (FCN) and U-Net. Several ablation studies are provided for detection, segmentation, and geometric primitive extraction. The proposed datasets are surgery videos of different patients. RESULTS In detection, ART-Net achieves 100.0% in both average precision and accuracy. In segmentation, it achieves 81.0% in mean Intersection over Union (mIoU) on the robotic EndoVis dataset (articulated tool), where it outperforms both FCN and U-Net, by 4.5pp and 2.9pp, respectively. It achieves 88.2% in mIoU on the remaining datasets (non-articulated tool). In geometric primitive extraction, ART-Net achieves 2.45∘ and 2.23∘ in mean Arc Length (mAL) error for the edge-lines and mid-line, respectively, and 9.3 pixels in mean Euclidean distance error for the tool-tip. Finally, in terms of 3D pose evaluated on animal data, our framework achieves 1.87 mm, 0.70 mm, and 4.80 mm mean absolute errors on the X,Y, and Z coordinates, respectively, and 5.94∘ angular error on the shaft orientation. It achieves 2.59 mm and 1.99 mm in mean and median location error of the tool head evaluated on patient data. CONCLUSIONS The proposed framework outperforms existing ones in detection and segmentation. Compared to separate networks, integrating the tasks in a single network preserves accuracy in detection and segmentation but substantially improves accuracy in geometric primitive extraction. Overall, our framework has similar or better accuracy in 3D pose estimation while largely improving robustness against the very challenging imaging conditions of laparoscopy. The source code of our framework and our annotated dataset will be made publicly available at https://github.com/kamruleee51/ART-Net.
Collapse
Affiliation(s)
- Md Kamrul Hasan
- EnCoV, Institut Pascal, UMR 6602 CNRS/Université Clermont-Auvergne, Clermont-Ferrand, France; Department of Electrical and Electronic Engineering, Khulna University of Engineering & Technology, Khulna 9203, Bangladesh.
| | - Lilian Calvet
- EnCoV, Institut Pascal, UMR 6602 CNRS/Université Clermont-Auvergne, Clermont-Ferrand, France
| | - Navid Rabbani
- EnCoV, Institut Pascal, UMR 6602 CNRS/Université Clermont-Auvergne, Clermont-Ferrand, France
| | - Adrien Bartoli
- EnCoV, Institut Pascal, UMR 6602 CNRS/Université Clermont-Auvergne, Clermont-Ferrand, France
| |
Collapse
|
11
|
He Y, Zhao B, Qi X, Li S, Yang Y, Hu Y. Automatic Surgical Field of View Control in Robot-Assisted Nasal Surgery. IEEE Robot Autom Lett 2021. [DOI: 10.1109/lra.2020.3039732] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
12
|
Fu Z, Jin Z, Zhang C, He Z, Zha Z, Hu C, Gan T, Yan Q, Wang P, Ye X. The Future of Endoscopic Navigation: A Review of Advanced Endoscopic Vision Technology. IEEE ACCESS 2021; 9:41144-41167. [DOI: 10.1109/access.2021.3065104] [Citation(s) in RCA: 16] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/04/2025]
|
13
|
Dehghani H, Sun Y, Cubrich L, Oleynikov D, Farritor S, Terry B. An Optimization-Based Algorithm for Trajectory Planning of an Under-Actuated Robotic Arm to Perform Autonomous Suturing. IEEE Trans Biomed Eng 2020; 68:1262-1272. [PMID: 32946377 DOI: 10.1109/tbme.2020.3024632] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
In single-port access surgeries, robot size is crucial due to the limited workspace. Thus, a robot may be designed under-actuated. Suturing, in contrast, is a complicated task and requires full actuation. This study aims to overcome this shortcoming by implementing an optimization-based algorithm for autonomous suturing for an under-actuated robot. The proposed algorithm approximates the ideal suturing trajectory by slightly reorienting the needle while deviating as little as possible from the ideal, full degree-of-freedom suturing case. The deviation of the path taken by a custom robot with respect to the ideal trajectory varies depending on the suturing starting location within the workspace as well as the needle size. A quantitative analysis reveals that in 13% of the investigated workspace, the accumulative deviation was less than 10 mm. In the remaining workspace, the accumulative deviation was less than 30 mm. Likewise, the accumulative deviation of a needle with a radius of 10 mm was 2.2 mm as opposed to 8 mm when the radius was 20 mm. The optimization-based algorithm maximized the accuracy of a four-DOF robot to perform a path-constrained trajectory and illustrates the accuracy-workspace correlation.
Collapse
|
14
|
Review of surgical robotic systems for keyhole and endoscopic procedures: state of the art and perspectives. Front Med 2020; 14:382-403. [DOI: 10.1007/s11684-020-0781-x] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2019] [Accepted: 03/05/2020] [Indexed: 02/06/2023]
|
15
|
Zhong F, Li P, Shi J, Wang Z, Wu J, Chan JYK, Leung N, Leung I, Tong MCF, Liu YH. Foot-Controlled Robot-Enabled EnDOscope Manipulator (FREEDOM) for Sinus Surgery: Design, Control, and Evaluation. IEEE Trans Biomed Eng 2019; 67:1530-1541. [PMID: 31494541 DOI: 10.1109/tbme.2019.2939557] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Despite successful clinical applications, teleoperated robotic surgical systems face particular limitations in the functional endoscopic sinus surgery (FESS) in terms of incompatible instrument dimensions and robot set-up. The endoscope remains manually handled by an assistant when the surgeon performs bimanual operations. This paper introduces the development of the Foot-controlled Robot-Enabled EnDOscope Manipulator (FREEDOM) designed for FESS. The system features clinical considerations that inform the design for providing reliable and safe endoscope positioning with minimal obstruction to the routine practice. The robot structure is modular and compact to ensure coaxial instrument manipulation through the nostril for manual procedures. To avoid rigid endoscope motions, a new compliant endoscope holder is proposed that passively limits the lens-tissue contact forces under collisions for patient-side protection. To facilitate hands-free endoscope manipulation that imposes minimal distractions to the surgeon, a foot-wearable interface is further designed to relieve the assistant's workload. The foot control method owns a short learning curve (mean 3.4 mins), and leads the task to be more ergonomic and surgeon-centered. Cadaver and clinical studies were both conducted to evaluate the surgical applicability of the FREEDOM to assist endoscope manipulation in FESS. The system was validated to be safe (IEC-60601-1) and easy for set up (mean 3.6 mins), from which the surgeon could perform various three-handed procedures alone in FESS without disrupting the routine practice.
Collapse
|
16
|
Colleoni E, Moccia S, Du X, De Momi E, Stoyanov D. Deep Learning Based Robotic Tool Detection and Articulation Estimation With Spatio-Temporal Layers. IEEE Robot Autom Lett 2019. [DOI: 10.1109/lra.2019.2917163] [Citation(s) in RCA: 57] [Impact Index Per Article: 9.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
17
|
Al Hajj H, Lamard M, Conze PH, Roychowdhury S, Hu X, Maršalkaitė G, Zisimopoulos O, Dedmari MA, Zhao F, Prellberg J, Sahu M, Galdran A, Araújo T, Vo DM, Panda C, Dahiya N, Kondo S, Bian Z, Vahdat A, Bialopetravičius J, Flouty E, Qiu C, Dill S, Mukhopadhyay A, Costa P, Aresta G, Ramamurthy S, Lee SW, Campilho A, Zachow S, Xia S, Conjeti S, Stoyanov D, Armaitis J, Heng PA, Macready WG, Cochener B, Quellec G. CATARACTS: Challenge on automatic tool annotation for cataRACT surgery. Med Image Anal 2018; 52:24-41. [PMID: 30468970 DOI: 10.1016/j.media.2018.11.008] [Citation(s) in RCA: 38] [Impact Index Per Article: 5.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2018] [Revised: 11/13/2018] [Accepted: 11/15/2018] [Indexed: 12/29/2022]
Abstract
Surgical tool detection is attracting increasing attention from the medical image analysis community. The goal generally is not to precisely locate tools in images, but rather to indicate which tools are being used by the surgeon at each instant. The main motivation for annotating tool usage is to design efficient solutions for surgical workflow analysis, with potential applications in report generation, surgical training and even real-time decision support. Most existing tool annotation algorithms focus on laparoscopic surgeries. However, with 19 million interventions per year, the most common surgical procedure in the world is cataract surgery. The CATARACTS challenge was organized in 2017 to evaluate tool annotation algorithms in the specific context of cataract surgery. It relies on more than nine hours of videos, from 50 cataract surgeries, in which the presence of 21 surgical tools was manually annotated by two experts. With 14 participating teams, this challenge can be considered a success. As might be expected, the submitted solutions are based on deep learning. This paper thoroughly evaluates these solutions: in particular, the quality of their annotations are compared to that of human interpretations. Next, lessons learnt from the differential analysis of these solutions are discussed. We expect that they will guide the design of efficient surgery monitoring tools in the near future.
Collapse
Affiliation(s)
| | - Mathieu Lamard
- Inserm, UMR 1101, Brest, F-29200, France; Univ Bretagne Occidentale, Brest, F-29200, France
| | - Pierre-Henri Conze
- Inserm, UMR 1101, Brest, F-29200, France; IMT Atlantique, LaTIM UMR 1101, UBL, Brest, F-29200, France
| | | | - Xiaowei Hu
- Dept. of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong, China
| | | | | | - Muneer Ahmad Dedmari
- Chair for Computer Aided Medical Procedures, Faculty of Informatics, Technical University of Munich, Garching b. Munich, 85748, Germany
| | - Fenqiang Zhao
- Key Laboratory of Biomedical Engineering of Ministry of Education, Zhejiang University, HangZhou, 310000, China
| | - Jonas Prellberg
- Dept. of Informatics, Carl von Ossietzky University, Oldenburg, 26129, Germany
| | - Manish Sahu
- Department of Visual Data Analysis, Zuse Institute Berlin, Berlin, 14195, Germany
| | - Adrian Galdran
- INESC TEC - Instituto de Engenharia de Sistemas e Computadores - Tecnologia e Ciência, Porto, 4200-465, Portugal
| | - Teresa Araújo
- Faculdade de Engenharia, Universidade do Porto, Porto, 4200-465, Portugal; INESC TEC - Instituto de Engenharia de Sistemas e Computadores - Tecnologia e Ciência, Porto, 4200-465, Portugal
| | - Duc My Vo
- Gachon University, 1342 Seongnamdaero, Sujeonggu, Seongnam, 13120, Korea
| | | | - Navdeep Dahiya
- Laboratory of Computational Computer Vision, Georgia Tech, Atlanta, GA, 30332, USA
| | | | | | - Arash Vahdat
- D-Wave Systems Inc., Burnaby, BC, V5G 4M9, Canada
| | | | | | - Chenhui Qiu
- Key Laboratory of Biomedical Engineering of Ministry of Education, Zhejiang University, HangZhou, 310000, China
| | - Sabrina Dill
- Department of Visual Data Analysis, Zuse Institute Berlin, Berlin, 14195, Germany
| | - Anirban Mukhopadhyay
- Department of Computer Science, Technische Universität Darmstadt, Darmstadt, 64283, Germany
| | - Pedro Costa
- INESC TEC - Instituto de Engenharia de Sistemas e Computadores - Tecnologia e Ciência, Porto, 4200-465, Portugal
| | - Guilherme Aresta
- Faculdade de Engenharia, Universidade do Porto, Porto, 4200-465, Portugal; INESC TEC - Instituto de Engenharia de Sistemas e Computadores - Tecnologia e Ciência, Porto, 4200-465, Portugal
| | - Senthil Ramamurthy
- Laboratory of Computational Computer Vision, Georgia Tech, Atlanta, GA, 30332, USA
| | - Sang-Woong Lee
- Gachon University, 1342 Seongnamdaero, Sujeonggu, Seongnam, 13120, Korea
| | - Aurélio Campilho
- Faculdade de Engenharia, Universidade do Porto, Porto, 4200-465, Portugal; INESC TEC - Instituto de Engenharia de Sistemas e Computadores - Tecnologia e Ciência, Porto, 4200-465, Portugal
| | - Stefan Zachow
- Department of Visual Data Analysis, Zuse Institute Berlin, Berlin, 14195, Germany
| | - Shunren Xia
- Key Laboratory of Biomedical Engineering of Ministry of Education, Zhejiang University, HangZhou, 310000, China
| | - Sailesh Conjeti
- Chair for Computer Aided Medical Procedures, Faculty of Informatics, Technical University of Munich, Garching b. Munich, 85748, Germany; German Center for Neurodegenrative Diseases (DZNE), Bonn, 53127, Germany
| | - Danail Stoyanov
- Digital Surgery Ltd, EC1V 2QY, London, UK; University College London, Gower Street, WC1E 6BT, London, UK
| | | | - Pheng-Ann Heng
- Dept. of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong, China
| | | | - Béatrice Cochener
- Inserm, UMR 1101, Brest, F-29200, France; Univ Bretagne Occidentale, Brest, F-29200, France; Service d'Ophtalmologie, CHRU Brest, Brest, F-29200, France
| | | |
Collapse
|
18
|
Hao R, Özgüner O, Çavuşoğlu MC. Vision-Based Surgical Tool Pose Estimation for the da Vinci ® Robotic Surgical System. PROCEEDINGS OF THE ... IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS. IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS 2018; 2018:1298-1305. [PMID: 31440395 PMCID: PMC6706092 DOI: 10.1109/iros.2018.8594471] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
This paper presents an approach to surgical tool tracking using stereo vision for the da Vinci® Surgical Robotic System. The proposed method is based on robot kinematics, computer vision techniques and Bayesian state estimation. The proposed method employs a silhouette rendering algorithm to create virtual images of the surgical tool by generating the silhouette of the defined tool geometry under the da Vinci® robot endoscopes. The virtual rendering method provides the tool representation in image form, which makes it possible to measure the distance between the rendered tool and real tool from endoscopic stereo image streams. Particle Filter algorithm employing the virtual rendering method is then used for surgical tool tracking. The tracking performance is evaluated on an actual da Vinci® surgical robotic system and a ROS/Gazebo-based simulation of the da Vinci® system.
Collapse
Affiliation(s)
- Ran Hao
- Department of Electrical Engineering and Computer Science, Case Western Reserve University, Cleveland, OH
| | - Orhan Özgüner
- Department of Electrical Engineering and Computer Science, Case Western Reserve University, Cleveland, OH
| | - M. Cenk Çavuşoğlu
- Department of Electrical Engineering and Computer Science, Case Western Reserve University, Cleveland, OH
| |
Collapse
|
19
|
Gruijthuijsen C, Dong L, Morel G, Poorten EV. Leveraging the Fulcrum Point in Robotic Minimally Invasive Surgery. IEEE Robot Autom Lett 2018. [DOI: 10.1109/lra.2018.2809495] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
20
|
Duflot LA, Reisenhofer R, Tamadazte B, Andreff N, Krupa A. Wavelet and shearlet-based image representations for visual servoing. Int J Rob Res 2018. [DOI: 10.1177/0278364918769739] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
A visual servoing scheme consists of a closed-loop control approach that uses visual information feedback to control the motion of a robotic system. Probably the most popular visual servoing method is image-based visual servoing (IBVS). This kind of method uses geometric visual features extracted from the image to design the control law. However, extracting, matching, and tracking geometric visual features over time significantly limits the versatility of visual servoing controllers in various industrial and medical applications, in particular for “low-structured” medical images, e.g. ultrasounds and optical coherence tomography modalities. To overcome the limits of conventional IBVS, one can consider novel visual servoing paradigms known as “ direct” or “ featureless” approaches. This paper deals with the development of a new generation of direct visual servoing methods in which the signal control inputs are the coefficients of a multiscale image representation. In particular, we consider the use of multiscale image representations that are based on discrete wavelet and shearlet transforms. Up to now, one of the main obstacles in the investigation of multiscale image representations for visual servoing schemes was the issue of obtaining an analytical formulation of the interaction matrix that links the variation of wavelet and shearlet coefficients to the spatial velocity of the camera and the robot. In this paper, we derive four direct visual servoing controllers: two that are based on subsampled respectively non-subsampled wavelet coefficients and two that are based on the coefficients of subsampled respectively non-subsampled discrete shearlet transforms. All proposed controllers were tested in both simulation and experimental scenarios (using a six-degree-of-freedom Cartesian robot in an eye-in-hand configuration). The objective of this paper is to provide an analysis of the respective strengths and weaknesses of wavelet- and shearlet-based visual servoing controllers.
Collapse
Affiliation(s)
- Lesley-Ann Duflot
- Université Rennes, Inria, CNRS, IRISA, Rennes, France
- FEMTO-ST, AS2M, Université Bourgogne Franche-Comté, Besançon, France
| | | | - Brahim Tamadazte
- FEMTO-ST, AS2M, Université Bourgogne Franche-Comté, Besançon, France
| | - Nicolas Andreff
- FEMTO-ST, AS2M, Université Bourgogne Franche-Comté, Besançon, France
| | | |
Collapse
|
21
|
Instrument detection and pose estimation with rigid part mixtures model in video-assisted surgeries. Med Image Anal 2018; 46:244-265. [DOI: 10.1016/j.media.2018.03.012] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2017] [Revised: 03/19/2018] [Accepted: 03/26/2018] [Indexed: 11/24/2022]
|
22
|
Yamaguchi T, Nakamura R. Laparoscopic training using a quantitative assessment and instructional system. Int J Comput Assist Radiol Surg 2018; 13:1453-1461. [DOI: 10.1007/s11548-018-1771-1] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2018] [Accepted: 04/16/2018] [Indexed: 01/13/2023]
|
23
|
Mansouri S, Farahmand F, Vossoughi G, Ghavidel AA, Rezayat M. Feasibility of infrared tracking of beating heart motion for robotic assisted beating heart surgery. Int J Med Robot 2017; 14. [DOI: 10.1002/rcs.1869] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2017] [Revised: 09/03/2017] [Accepted: 09/05/2017] [Indexed: 01/23/2023]
Affiliation(s)
- Saeed Mansouri
- Department of Mechanical Engineering; Sharif University of Technology; Tehran Iran
| | - Farzam Farahmand
- Department of Mechanical Engineering; Sharif University of Technology; Tehran Iran
- RCBTR; Tehran University of Medical Sciences; Tehran Iran
| | - Gholamreza Vossoughi
- Department of Mechanical Engineering; Sharif University of Technology; Tehran Iran
| | - Alireza Alizadeh Ghavidel
- Heart Valve Disease Research Center, Rajaie Cardiovascular Medical and Research Center; Iran University of Medical Sciences; Tehran Iran
| | | |
Collapse
|
24
|
Zhao Z, Voros S, Weng Y, Chang F, Li R. Tracking-by-detection of surgical instruments in minimally invasive surgery via the convolutional neural network deep learning-based method. Comput Assist Surg (Abingdon) 2017; 22:26-35. [DOI: 10.1080/24699322.2017.1378777] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022] Open
Affiliation(s)
- Zijian Zhao
- School of Control Science and Engineering, Shandong University, Jinan, China
| | - Sandrine Voros
- CNRS, INSERM, TIMC-IMAG, University Grenoble-Alpes, Grenoble, France
| | - Ying Weng
- School of Computer Science, Bangor University, Bangor, UK
| | - Faliang Chang
- School of Control Science and Engineering, Shandong University, Jinan, China
| | - Ruijian Li
- Department of cardiology, Qilu Hospital of Shandong University, Jinan, China
| |
Collapse
|
25
|
Wang Z, Lee SC, Zhong F, Navarro-Alarcon D, Liu YH, Deguet A, Kazanzides P, Taylor RH. Image-Based Trajectory Tracking Control of 4-DoF Laparoscopic Instruments Using a Rotation Distinguishing Marker. IEEE Robot Autom Lett 2017. [DOI: 10.1109/lra.2017.2676350] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
26
|
Sarikaya D, Corso JJ, Guru KA. Detection and Localization of Robotic Tools in Robot-Assisted Surgery Videos Using Deep Neural Networks for Region Proposal and Detection. IEEE TRANSACTIONS ON MEDICAL IMAGING 2017; 36:1542-1549. [PMID: 28186883 DOI: 10.1109/tmi.2017.2665671] [Citation(s) in RCA: 97] [Impact Index Per Article: 12.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Video understanding of robot-assisted surgery (RAS) videos is an active research area. Modeling the gestures and skill level of surgeons presents an interesting problem. The insights drawn may be applied in effective skill acquisition, objective skill assessment, real-time feedback, and human-robot collaborative surgeries. We propose a solution to the tool detection and localization open problem in RAS video understanding, using a strictly computer vision approach and the recent advances of deep learning. We propose an architecture using multimodal convolutional neural networks for fast detection and localization of tools in RAS videos. To the best of our knowledge, this approach will be the first to incorporate deep neural networks for tool detection and localization in RAS videos. Our architecture applies a region proposal network (RPN) and a multimodal two stream convolutional network for object detection to jointly predict objectness and localization on a fusion of image and temporal motion cues. Our results with an average precision of 91% and a mean computation time of 0.1 s per test frame detection indicate that our study is superior to conventionally used methods for medical imaging while also emphasizing the benefits of using RPN for precision and efficiency. We also introduce a new data set, ATLAS Dione, for RAS video understanding. Our data set provides video data of ten surgeons from Roswell Park Cancer Institute, Buffalo, NY, USA, performing six different surgical tasks on the daVinci Surgical System (dVSS) with annotations of robotic tools per frame.
Collapse
|
27
|
Trujano MA, Garrido R, Soria A. Robust Visual Control of Parallel Robots under Uncertain Camera Orientation. INT J ADV ROBOT SYST 2017. [DOI: 10.5772/51743] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022] Open
Abstract
This work presents a stability analysis and experimental assessment of a visual control algorithm applied to a redundant planar parallel robot under uncertainty in relation to camera orientation. The key feature of the analysis is a strict Lyapunov function that allows the conclusion of asymptotic stability without invoking the Barbashin-Krassovsky-LaSalle invariance theorem. The controller does not rely on velocity measurements and has a structure similar to a classic Proportional Derivative control algorithm. Experiments in a laboratory prototype show that uncertainty in camera orientation does not significantly degrade closed-loop performance.
Collapse
Affiliation(s)
- Miguel A. Trujano
- Departamento de Control Automático, CINVESTAV-IPN, Av. IPN 2508 San Pedro Zacatenco, Mexico
| | - Rubén Garrido
- Departamento de Control Automático, CINVESTAV-IPN, Av. IPN 2508 San Pedro Zacatenco, Mexico
| | - Alberto Soria
- Departamento de Control Automático, CINVESTAV-IPN, Av. IPN 2508 San Pedro Zacatenco, Mexico
| |
Collapse
|
28
|
Sen HT, Bell MAL, Zhang Y, Ding K, Boctor E, Wong J, Iordachita I, Kazanzides P. System Integration and In Vivo Testing of a Robot for Ultrasound Guidance and Monitoring During Radiotherapy. IEEE Trans Biomed Eng 2016; 64:1608-1618. [PMID: 28113225 DOI: 10.1109/tbme.2016.2612229] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/28/2022]
Abstract
We are developing a cooperatively controlled robot system for image-guided radiation therapy (IGRT) in which a clinician and robot share control of a 3-D ultrasound (US) probe. IGRT involves two main steps: 1) planning/simulation and 2) treatment delivery. The goals of the system are to provide guidance for patient setup and real-time target monitoring during fractionated radiotherapy of soft tissue targets, especially in the upper abdomen. To compensate for soft tissue deformations created by the probe, we present a novel workflow where the robot holds the US probe on the patient during acquisition of the planning computerized tomography image, thereby ensuring that planning is performed on the deformed tissue. The robot system introduces constraints (virtual fixtures) to help to produce consistent soft tissue deformation between simulation and treatment days, based on the robot position, contact force, and reference US image recorded during simulation. This paper presents the system integration and the proposed clinical workflow, validated by an in vivo canine study. The results show that the virtual fixtures enable the clinician to deviate from the recorded position to better reproduce the reference US image, which correlates with more consistent soft tissue deformation and the possibility for more accurate patient setup and radiation delivery.
Collapse
|
29
|
Bouget D, Allan M, Stoyanov D, Jannin P. Vision-based and marker-less surgical tool detection and tracking: a review of the literature. Med Image Anal 2016; 35:633-654. [PMID: 27744253 DOI: 10.1016/j.media.2016.09.003] [Citation(s) in RCA: 112] [Impact Index Per Article: 12.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2016] [Revised: 06/26/2016] [Accepted: 09/05/2016] [Indexed: 11/16/2022]
Abstract
In recent years, tremendous progress has been made in surgical practice for example with Minimally Invasive Surgery (MIS). To overcome challenges coming from deported eye-to-hand manipulation, robotic and computer-assisted systems have been developed. Having real-time knowledge of the pose of surgical tools with respect to the surgical camera and underlying anatomy is a key ingredient for such systems. In this paper, we present a review of the literature dealing with vision-based and marker-less surgical tool detection. This paper includes three primary contributions: (1) identification and analysis of data-sets used for developing and testing detection algorithms, (2) in-depth comparison of surgical tool detection methods from the feature extraction process to the model learning strategy and highlight existing shortcomings, and (3) analysis of validation techniques employed to obtain detection performance results and establish comparison between surgical tool detectors. The papers included in the review were selected through PubMed and Google Scholar searches using the keywords: "surgical tool detection", "surgical tool tracking", "surgical instrument detection" and "surgical instrument tracking" limiting results to the year range 2000 2015. Our study shows that despite significant progress over the years, the lack of established surgical tool data-sets, and reference format for performance assessment and method ranking is preventing faster improvement.
Collapse
Affiliation(s)
- David Bouget
- Medicis team, INSERM U1099, Université de Rennes 1 LTSI, 35000 Rennes, France.
| | - Max Allan
- Center for Medical Image Computing. University College London, WC1E 6BT London, United Kingdom.
| | - Danail Stoyanov
- Center for Medical Image Computing. University College London, WC1E 6BT London, United Kingdom.
| | - Pierre Jannin
- Medicis team, INSERM U1099, Université de Rennes 1 LTSI, 35000 Rennes, France.
| |
Collapse
|
30
|
Voros S, Long JA, Cinquin P. Automatic Detection of Instruments in Laparoscopic Images: A First Step Towards High-level Command of Robotic Endoscopic Holders. Int J Rob Res 2016. [DOI: 10.1177/0278364907083395] [Citation(s) in RCA: 70] [Impact Index Per Article: 7.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
The tracking of surgical instruments o fers interesting possibilities for the development of high-level commands for robotic camera holders in laparoscopic surgery. We have developed a new method to detect instruments in laparoscopic images which uses information on the 3D position of the insertion point of an instrument into the abdominal cavity. This information strongly constrains the search for the instrument in each endoscopic image. Hence, the instrument can be detected in near real-time using shape considerations. Early results on laparoscopic images show that the method is rapid and robust in the presence of partial occlusion and smoke. Our first experiment on a cadaver validates our approach and shows encouraging results.
Collapse
Affiliation(s)
- Sandrine Voros
- Université J. Fourier, Laboratoire TIMC-IMAG, CNRS UMR 5525, INSERM, IFR 130, F-38000 Grenoble, France
| | | | - Philippe Cinquin
- Université J. Fourier, Laboratoire TIMC-IMAG CNRS, UMR 5525, INSERM, IFR 130 F-38000 Grenoble, France
| |
Collapse
|
31
|
Amini Khoiy K, Mirbagheri A, Farahmand F. Automatic tracking of laparoscopic instruments for autonomous control of a cameraman robot. MINIM INVASIV THER 2016; 25:121-8. [DOI: 10.3109/13645706.2016.1141101] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
|
32
|
Bouget D, Benenson R, Omran M, Riffaud L, Schiele B, Jannin P. Detecting Surgical Tools by Modelling Local Appearance and Global Shape. IEEE TRANSACTIONS ON MEDICAL IMAGING 2015; 34:2603-2617. [PMID: 26625340 DOI: 10.1109/tmi.2015.2450831] [Citation(s) in RCA: 54] [Impact Index Per Article: 5.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
Detecting tools in surgical videos is an important ingredient for context-aware computer-assisted surgical systems. To this end, we present a new surgical tool detection dataset and a method for joint tool detection and pose estimation in 2d images. Our two-stage pipeline is data-driven and relaxes strong assumptions made by previous works regarding the geometry, number, and position of tools in the image. The first stage classifies each pixel based on local appearance only, while the second stage evaluates a tool-specific shape template to enforce global shape. Both local appearance and global shape are learned from training data. Our method is validated on a new surgical tool dataset of 2 476 images from neurosurgical microscopes, which is made freely available. It improves over existing datasets in size, diversity and detail of annotation. We show that our method significantly improves over competitive baselines from the computer vision field. We achieve 15% detection miss-rate at 10(-1) false positives per image (for the suction tube) over our surgical tool dataset. Results indicate that performing semantic labelling as an intermediate task is key for high quality detection.
Collapse
|
33
|
Kassahun Y, Yu B, Tibebu AT, Stoyanov D, Giannarou S, Metzen JH, Vander Poorten E. Surgical robotics beyond enhanced dexterity instrumentation: a survey of machine learning techniques and their role in intelligent and autonomous surgical actions. Int J Comput Assist Radiol Surg 2015; 11:553-68. [PMID: 26450107 DOI: 10.1007/s11548-015-1305-z] [Citation(s) in RCA: 96] [Impact Index Per Article: 9.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2015] [Accepted: 09/21/2015] [Indexed: 02/06/2023]
Abstract
PURPOSE Advances in technology and computing play an increasingly important role in the evolution of modern surgical techniques and paradigms. This article reviews the current role of machine learning (ML) techniques in the context of surgery with a focus on surgical robotics (SR). Also, we provide a perspective on the future possibilities for enhancing the effectiveness of procedures by integrating ML in the operating room. METHODS The review is focused on ML techniques directly applied to surgery, surgical robotics, surgical training and assessment. The widespread use of ML methods in diagnosis and medical image computing is beyond the scope of the review. Searches were performed on PubMed and IEEE Explore using combinations of keywords: ML, surgery, robotics, surgical and medical robotics, skill learning, skill analysis and learning to perceive. RESULTS Studies making use of ML methods in the context of surgery are increasingly being reported. In particular, there is an increasing interest in using ML for developing tools to understand and model surgical skill and competence or to extract surgical workflow. Many researchers begin to integrate this understanding into the control of recent surgical robots and devices. CONCLUSION ML is an expanding field. It is popular as it allows efficient processing of vast amounts of data for interpreting and real-time decision making. Already widely used in imaging and diagnosis, it is believed that ML will also play an important role in surgery and interventional treatments. In particular, ML could become a game changer into the conception of cognitive surgical robots. Such robots endowed with cognitive skills would assist the surgical team also on a cognitive level, such as possibly lowering the mental load of the team. For example, ML could help extracting surgical skill, learned through demonstration by human experts, and could transfer this to robotic skills. Such intelligent surgical assistance would significantly surpass the state of the art in surgical robotics. Current devices possess no intelligence whatsoever and are merely advanced and expensive instruments.
Collapse
Affiliation(s)
- Yohannes Kassahun
- Robotics Innovation Center, German Research Center for Artificial Intelligence, Robert-Hooke-Str. 1, 28359, Bremen, Germany.
| | - Bingbin Yu
- Faculty 3 - Mathematics and Computer Science, University of Bremen, Robert-Hooke-Str. 1, 28359, Bremen, Germany
| | - Abraham Temesgen Tibebu
- Faculty 3 - Mathematics and Computer Science, University of Bremen, Robert-Hooke-Str. 1, 28359, Bremen, Germany
| | - Danail Stoyanov
- Centre for Medical Image Computing, Department of Computer Science, University College London, London, UK
| | | | - Jan Hendrik Metzen
- Faculty 3 - Mathematics and Computer Science, University of Bremen, Robert-Hooke-Str. 1, 28359, Bremen, Germany
| | - Emmanuel Vander Poorten
- Department of Mechanical Engineering, University of Leuven, Celestijnenlaan 300B, 3001, Heverlee, Belgium
| |
Collapse
|
34
|
Vision-based real-time position control of a semi-automated system for robot-assisted joint fracture surgery. Int J Comput Assist Radiol Surg 2015; 11:437-55. [PMID: 26429787 DOI: 10.1007/s11548-015-1296-9] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2015] [Accepted: 09/10/2015] [Indexed: 01/19/2023]
Abstract
PURPOSE Joint fracture surgery quality can be improved by robotic system with high-accuracy and high-repeatability fracture fragment manipulation. A new real-time vision-based system for fragment manipulation during robot-assisted fracture surgery was developed and tested. METHODS The control strategy was accomplished by merging fast open-loop control with vision-based control. This two-phase process is designed to eliminate the open-loop positioning errors by closing the control loop using visual feedback provided by an optical tracking system. Evaluation of the control system accuracy was performed using robot positioning trials, and fracture reduction accuracy was tested in trials on ex vivo porcine model. RESULTS The system resulted in high fracture reduction reliability with a reduction accuracy of 0.09 mm (translations) and of [Formula: see text] (rotations), maximum observed errors in the order of 0.12 mm (translations) and of [Formula: see text] (rotations), and a reduction repeatability of 0.02 mm and [Formula: see text]. CONCLUSIONS The proposed vision-based system was shown to be effective and suitable for real joint fracture surgical procedures, contributing a potential improvement of their quality.
Collapse
|
35
|
Computer-simulated biopsy marking system for endoscopic surveillance of gastric lesions: a pilot study. BIOMED RESEARCH INTERNATIONAL 2015; 2015:197270. [PMID: 25954747 PMCID: PMC4411451 DOI: 10.1155/2015/197270] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/19/2014] [Revised: 08/11/2014] [Accepted: 08/15/2014] [Indexed: 11/29/2022]
Abstract
Endoscopic tattoo with India ink injection for surveillance of premalignant gastric lesions is technically cumbersome and may not be durable. The aim of the study is to evaluate the accuracy of a novel, computer-simulated biopsy marking system (CSBMS) developed for the endoscopic marking of gastric lesions. Twenty-five patients with history of gastric intestinal metaplasia received both CSBMS-guided marking and India ink injection in five points in the stomach at index endoscopy. A second endoscopy was performed at three months. Primary outcome was accuracy of CSBMS (distance between CSBMS probe-guided site and tattoo site measured by CSBMS). The mean accuracy of CSBMS at angularis was 5.3 ± 2.2 mm, antral lesser curvature 5.7 ± 1.4 mm, antral greater curvature 6.1 ± 1.1 mm, antral anterior wall 6.9 ± 1.6 mm, and antral posterior wall 6.9 ± 1.6 mm. CSBMS (2.3 ± 0.9 versus 12.5 ± 4.6 seconds; P = 0.02) required less procedure time compared to endoscopic tattooing. No adverse events were encountered. CSBMS accurately identified previously marked gastric sites by endoscopic tattooing within 1 cm on follow-up endoscopy.
Collapse
|
36
|
Okada S, Shimada J, Ito K, Kato D. A touch panel surgical navigation system with automatic depth perception. Int J Comput Assist Radiol Surg 2014; 10:243-51. [PMID: 24906296 DOI: 10.1007/s11548-014-1080-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2014] [Accepted: 05/21/2014] [Indexed: 12/01/2022]
Abstract
PURPOSE A touch panel navigation system may be used to enhance endoscopic surgery, especially for cauterization. We developed and tested the in vitro performance of a new touch panel navigation (TPN) system. METHODS This TPN system uses finger motion trajectories on a touch panel to control an argon plasma coagulation (APC) attached to a robot arm. Thermal papers with printed figures were soaked in saline for repeated recording and analysis of cauterized trajectory. A novice and an expert surgeon traced squares and circles displayed on the touch panel and cauterized them using the APC. Sixteen novices and eight experts cauterized squares and circles using both conventional endoscopic and TPN procedures. Six novices cauterized arcs using the endoscopic and TPN procedures 20 times a day for 5 consecutive days. RESULTS For square shapes, the offset was 5.5 mm with differences between the novice and the expert at 2 of 16 points. For circles, the offset was 5.0 mm and did not differ at any point. Task completion time for the TPN procedure was significantly longer than that for the endoscopic procedure for both squares and circles. For squares, the distance from the target for the TPN procedure was significantly smaller than that for the endoscopic procedure. For circles, the distance did not differ. There was no difference in task completion time and distance between the novices and the experts. Task completion time and distance improved significantly for the endoscopic procedure but not for the TPN procedure. CONCLUSIONS A new TPN system enabled the surgeons to accomplish continuous 3D positioning of the surgical device with automatic depth perception using finger tracing on a 2D monitor. This technology is promising for application in surgical procedures that require precise control of cauterization.
Collapse
Affiliation(s)
- Satoru Okada
- Division of Chest Surgery, Department of Surgery, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, 465 Kajii-cho, Kawaramachi-Hirokoji, Kamigyo-ku, Kyoto , 602-8566, Japan
| | | | | | | |
Collapse
|
37
|
A vision-based system for fast and accurate laser scanning in robot-assisted phonomicrosurgery. Int J Comput Assist Radiol Surg 2014; 10:217-29. [DOI: 10.1007/s11548-014-1078-9] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2014] [Accepted: 05/21/2014] [Indexed: 10/25/2022]
|
38
|
Sánchez-Margallo JA, Sánchez-Margallo FM, Pagador Carrasco JB, Oropesa García I, Gómez Aguilera EJ, Moreno del Pozo J. Usefulness of an Optical Tracking System in Laparoscopic Surgery for Motor Skills Assessment. ACTA ACUST UNITED AC 2014. [DOI: 10.1016/j.cireng.2013.01.012] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/22/2023]
|
39
|
Sánchez-Margallo JA, Sánchez-Margallo FM, Pagador Carrasco JB, Oropesa García I, Gómez Aguilera EJ, Moreno del Pozo J. Utilidad de un sistema de seguimiento óptico de instrumental en cirugía laparoscópica para evaluación de destrezas motoras. Cir Esp 2014; 92:421-8. [DOI: 10.1016/j.ciresp.2013.01.006] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2012] [Revised: 01/04/2013] [Accepted: 01/07/2013] [Indexed: 01/22/2023]
|
40
|
Rosa B, Gruijthuijsen C, Van Cleynenbreugel B, Sloten JV, Reynaerts D, Poorten EV. Estimation of optimal pivot point for remote center of motion alignment in surgery. Int J Comput Assist Radiol Surg 2014; 10:205-15. [PMID: 24830535 DOI: 10.1007/s11548-014-1071-3] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2014] [Accepted: 05/02/2014] [Indexed: 11/24/2022]
Affiliation(s)
- Benoît Rosa
- KU Leuven, Department of Mechanical Engineering, 3001 , Leuven, Belgium,
| | | | | | | | | | | |
Collapse
|
41
|
Abstract
In this paper, we present an appearance learning approach which is used to detect and track surgical robotic tools in laparoscopic sequences. By training a robust visual feature descriptor on low-level landmark features, we build a framework for fusing robot kinematics and 3D visual observations to track surgical tools over long periods of time across various types of environment. We demonstrate 3D tracking on multiple types of tool (with different overall appearances) as well as multiple tools simultaneously. We present experimental results using the da Vinci® surgical robot using a combination of both ex-vivo and in-vivo environments.
Collapse
Affiliation(s)
- Austin Reiter
- Department of Computer Science, Columbia University, USA
| | - Peter K Allen
- Department of Computer Science, Columbia University, USA
| | - Tao Zhao
- Intuitive Surgical, Inc., CA, USA
| |
Collapse
|
42
|
Bauzano E, Garcia-Morales I, del Saz-Orozco P, Fraile JC, Muñoz VF. A minimally invasive surgery robotic assistant for HALS-SILS techniques. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2013; 112:272-283. [PMID: 23566709 DOI: 10.1016/j.cmpb.2013.01.017] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/01/2012] [Accepted: 01/30/2013] [Indexed: 06/02/2023]
Abstract
This paper is focused in the design and implementation of a robotic surgical motion controller. The proposed control scheme addresses the issues related to the application of a robot assistant in novel surgical scenario, which combines hand assisted laparoscopic surgery (HALS) with the single incision laparoscopic surgery (SILS) techniques. It is designed for collaborating with the surgeon in a natural way, by performing autonomous movements, in order to assist the surgeon during a surgical maneuver. In this way, it is implemented a hierarchical architecture which includes an upper auto-guide velocity planner connected to a low-level force feedback controller. The first one, based on a behavior approach, computes a collision free trajectory of the surgical instrument tip, held by the robot, for reaching a goal location inside of the abdominal cavity. On the other hand, the force feedback controller uses this trajectory for performing the instrument displacement by taking into account the holonomic movement constraints introduced by the fulcrum point. The aim of this controller is positioning the surgical instrument by minimizing the forces exerted over the abdominal wall due to the fulcrum location uncertainty. The overall system has been integrated in the control architecture of the surgical assistant CISOBOT, designed and developed at the University of Malaga. The whole architecture performance has been tested by means of in vitro trials.
Collapse
Affiliation(s)
- E Bauzano
- Department of System Engineering and Automation, University of Malaga, Edificio de Institutos Universitarios, Labs. 9-10, Severo Ochoa 4, 29590 Malaga, Spain
| | | | | | | | | |
Collapse
|
43
|
Azizian M, Khoshnam M, Najmaei N, Patel RV. Visual servoing in medical robotics: a survey. Part I: endoscopic and direct vision imaging - techniques and applications. Int J Med Robot 2013; 10:263-74. [PMID: 24106103 DOI: 10.1002/rcs.1531] [Citation(s) in RCA: 48] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/03/2013] [Revised: 08/05/2013] [Accepted: 08/08/2013] [Indexed: 12/20/2022]
Abstract
BACKGROUND Intra-operative imaging is widely used to provide visual feedback to a clinician when he/she performs a procedure. In visual servoing, surgical instruments and parts of tissue/body are tracked by processing the acquired images. This information is then used within a control loop to manoeuvre a robotic manipulator during a procedure. METHODS A comprehensive search of electronic databases was completed for the period 2000-2013 to provide a survey of the visual servoing applications in medical robotics. The focus is on medical applications where image-based tracking is used for closed-loop control of a robotic system. RESULTS Detailed classification and comparative study of various contributions in visual servoing using endoscopic or direct visual images are presented and summarized in tables and diagrams. CONCLUSION The main challenges in using visual servoing for medical robotic applications are identified and potential future directions are suggested. 'Supervised automation of medical robotics' is found to be a major trend in this field.
Collapse
|
44
|
Allan M, Ourselin S, Thompson S, Hawkes DJ, Kelly J, Stoyanov D. Toward detection and localization of instruments in minimally invasive surgery. IEEE Trans Biomed Eng 2012. [PMID: 23192482 DOI: 10.1109/tbme.2012.2229278] [Citation(s) in RCA: 54] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Methods for detecting and localizing surgical instruments in laparoscopic images are an important element of advanced robotic and computer-assisted interventions. Robotic joint encoders and sensors integrated or mounted on the instrument can provide information about the tool's position, but this often has inaccuracy when transferred to the surgeon's point of view. Vision sensors are currently a promising approach for determining the position of instruments in the coordinate frame of the surgical camera. In this study, we propose a vision algorithm for localizing the instrument's pose in 3-D leaving only rotation in the axis of the tool's shaft as an ambiguity. We propose a probabilistic supervised classification method to detect pixels in laparoscopic images that belong to surgical tools. We then use the classifier output to initialize an energy minimization algorithm for estimating the pose of a prior 3-D model of the instrument within a level set framework. We show that the proposed method is robust against noise using simulated data and we perform quantitative validation of the algorithm compared to ground truth obtained using an optical tracker. Finally, we demonstrate the practical application of the technique on in vivo data from minimally invasive surgery with traditional laparoscopic and robotic instruments.
Collapse
Affiliation(s)
- Max Allan
- Centre for Medical Image Computing and the Department of Computer Science, University College London, London, UK.
| | | | | | | | | | | |
Collapse
|
45
|
EVA: laparoscopic instrument tracking based on Endoscopic Video Analysis for psychomotor skills assessment. Surg Endosc 2012; 27:1029-39. [PMID: 23052495 DOI: 10.1007/s00464-012-2513-z] [Citation(s) in RCA: 57] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/17/2012] [Accepted: 07/27/2012] [Indexed: 10/27/2022]
Abstract
INTRODUCTION The EVA (Endoscopic Video Analysis) tracking system is a new system for extracting motions of laparoscopic instruments based on nonobtrusive video tracking. The feasibility of using EVA in laparoscopic settings has been tested in a box trainer setup. METHODS EVA makes use of an algorithm that employs information of the laparoscopic instrument's shaft edges in the image, the instrument's insertion point, and the camera's optical center to track the three-dimensional position of the instrument tip. A validation study of EVA comprised a comparison of the measurements achieved with EVA and the TrEndo tracking system. To this end, 42 participants (16 novices, 22 residents, and 4 experts) were asked to perform a peg transfer task in a box trainer. Ten motion-based metrics were used to assess their performance. RESULTS Construct validation of the EVA has been obtained for seven motion-based metrics. Concurrent validation revealed that there is a strong correlation between the results obtained by EVA and the TrEndo for metrics, such as path length (ρ = 0.97), average speed (ρ = 0.94), or economy of volume (ρ = 0.85), proving the viability of EVA. CONCLUSIONS EVA has been successfully validated in a box trainer setup, showing the potential of endoscopic video analysis to assess laparoscopic psychomotor skills. The results encourage further implementation of video tracking in training setups and image-guided surgery.
Collapse
|
46
|
Najmaei N, Mostafavi K, Shahbazi S, Azizian M. Image-guided techniques in renal and hepatic interventions. Int J Med Robot 2012; 9:379-95. [DOI: 10.1002/rcs.1443] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/30/2012] [Indexed: 12/24/2022]
Affiliation(s)
- Nima Najmaei
- Canadian Surgical Technologies and Advanced Robotics (CSTAR); London Health Science Center; London ON Canada
- Department of Electrical and Computer Engineering; University of Western Ontario; London ON Canada
| | - Kamal Mostafavi
- Department of Mechanical Engineering; University of Western Ontario; London ON Canada
| | - Sahar Shahbazi
- Department of Electrical and Computer Engineering; University of Western Ontario; London ON Canada
| | - Mahdi Azizian
- Sheikh Zayed Institute for Pediatric Surgical Innovation; Children's National Medical Center; Washington DC USA
| |
Collapse
|
47
|
Climent J, Hexsel RA. Particle filtering in the Hough space for instrument tracking. Comput Biol Med 2012; 42:614-23. [DOI: 10.1016/j.compbiomed.2012.02.007] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2009] [Revised: 09/08/2011] [Accepted: 02/23/2012] [Indexed: 10/28/2022]
|
48
|
Krupa A, Morel G, de Mathelin M. Achieving high-precision laparoscopic manipulation through adaptive force control. Adv Robot 2012. [DOI: 10.1163/1568553042225769] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
49
|
Allain B, Hu M, Lovat LB, Cook RJ, Vercauteren T, Ourselin S, Hawkes DJ. Re-localisation of a biopsy site in endoscopic images and characterisation of its uncertainty. Med Image Anal 2012; 16:482-96. [DOI: 10.1016/j.media.2011.11.005] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2010] [Revised: 11/22/2011] [Accepted: 11/22/2011] [Indexed: 11/30/2022]
|
50
|
Stoyanov D. Surgical vision. Ann Biomed Eng 2011; 40:332-45. [PMID: 22012086 DOI: 10.1007/s10439-011-0441-z] [Citation(s) in RCA: 55] [Impact Index Per Article: 3.9] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2011] [Accepted: 10/07/2011] [Indexed: 10/16/2022]
Abstract
The emergence of Minimal Access Surgery (MAS) as a paradigm in modern healthcare treatment has created new challenges and opportunities for automated image understanding and computer vision. In MAS, images recovered from inside the body using specialized devices are used to visualize and operate on the surgical site but they can also be used to computationally infer in vivo 3D tissue surface shape, soft-tissue morphology, and surgical instrument motion. This information is important for facilitating in vivo biophotonic imaging modalities where the interaction between light and tissue is used to infer the structural and functional properties of the tissue. This article provides a review of the literature for computer vision and image understanding techniques applied to MAS images. The focus of this article is to elucidate a perspective on how computer vision techniques can be used to support and enhance the capabilities of biophotonic imaging modalities during surgery. Note that while MAS encompasses a variety of surgical specializations this review does not involve procedures performed in the interventional suite. The review has been carried out based on searches in the PubMed and IEEE databases using the article's keywords.
Collapse
Affiliation(s)
- Danail Stoyanov
- Center for Medical Image Computing, University College London, London, WC1 2BT, UK.
| |
Collapse
|