1
|
Tomita H, Ienaga N, Kajita H, Hayashida T, Sugimoto M. An analysis on the effect of body tissues and surgical tools on workflow recognition in first person surgical videos. Int J Comput Assist Radiol Surg 2024; 19:2195-2202. [PMID: 38411780 PMCID: PMC11541397 DOI: 10.1007/s11548-024-03074-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2023] [Accepted: 02/09/2024] [Indexed: 02/28/2024]
Abstract
PURPOSE Analysis of operative fields is expected to aid in estimating procedural workflow and evaluating surgeons' procedural skills by considering the temporal transitions during the progression of the surgery. This study aims to propose an automatic recognition system for the procedural workflow by employing machine learning techniques to identify and distinguish elements in the operative field, including body tissues such as fat, muscle, and dermis, along with surgical tools. METHODS We conducted annotations on approximately 908 first-person-view images of breast surgery to facilitate segmentation. The annotated images were used to train a pixel-level classifier based on Mask R-CNN. To assess the impact on procedural workflow recognition, we annotated an additional 43,007 images. The network, structured on the Transformer architecture, was then trained with surgical images incorporating masks for body tissues and surgical tools. RESULTS The instance segmentation of each body tissue in the segmentation phase provided insights into the trend of area transitions for each tissue. Simultaneously, the spatial features of the surgical tools were effectively captured. In regard to the accuracy of procedural workflow recognition, accounting for body tissues led to an average improvement of 3 % over the baseline. Furthermore, the inclusion of surgical tools yielded an additional increase in accuracy by 4 % compared to the baseline. CONCLUSION In this study, we revealed the contribution of the temporal transition of the body tissues and surgical tools spatial features to recognize procedural workflow in first-person-view surgical videos. Body tissues, especially in open surgery, can be a crucial element. This study suggests that further improvements can be achieved by accurately identifying surgical tools specific to each procedural workflow step.
Collapse
Affiliation(s)
- Hisako Tomita
- Graduate School of Science and Technology, Keio University, Yokohama, 2238522, Japan.
| | - Naoto Ienaga
- Institute of Systems and Information Engineering, University of Tsukuba, Tsukuba, 3058573, Japan
| | - Hiroki Kajita
- Department of Plastic and Reconstructive Surgery, Keio University School of Medicine, Tokyo, 1608582, Japan
| | - Tetsu Hayashida
- Department of Surgery, Keio University School of Medicine, Tokyo, 1608582, Japan
| | - Maki Sugimoto
- Graduate School of Science and Technology, Keio University, Yokohama, 2238522, Japan
| |
Collapse
|
2
|
Zuckerman I, Werner N, Kouchly J, Huston E, DiMarco S, DiMusto P, Laufer S. Depth over RGB: automatic evaluation of open surgery skills using depth camera. Int J Comput Assist Radiol Surg 2024; 19:1349-1357. [PMID: 38748053 PMCID: PMC11230951 DOI: 10.1007/s11548-024-03158-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2024] [Accepted: 04/22/2024] [Indexed: 07/10/2024]
Abstract
PURPOSE In this paper, we present a novel approach to the automatic evaluation of open surgery skills using depth cameras. This work is intended to show that depth cameras achieve similar results to RGB cameras, which is the common method in the automatic evaluation of open surgery skills. Moreover, depth cameras offer advantages such as robustness to lighting variations, camera positioning, simplified data compression, and enhanced privacy, making them a promising alternative to RGB cameras. METHODS Experts and novice surgeons completed two simulators of open suturing. We focused on hand and tool detection and action segmentation in suturing procedures. YOLOv8 was used for tool detection in RGB and depth videos. Furthermore, UVAST and MSTCN++ were used for action segmentation. Our study includes the collection and annotation of a dataset recorded with Azure Kinect. RESULTS We demonstrated that using depth cameras in object detection and action segmentation achieves comparable results to RGB cameras. Furthermore, we analyzed 3D hand path length, revealing significant differences between experts and novice surgeons, emphasizing the potential of depth cameras in capturing surgical skills. We also investigated the influence of camera angles on measurement accuracy, highlighting the advantages of 3D cameras in providing a more accurate representation of hand movements. CONCLUSION Our research contributes to advancing the field of surgical skill assessment by leveraging depth cameras for more reliable and privacy evaluations. The findings suggest that depth cameras can be valuable in assessing surgical skills and provide a foundation for future research in this area.
Collapse
Affiliation(s)
- Ido Zuckerman
- Faculty of Data and Decision Sciences, Technion - Israel Institute of Technology, Haifa, 3200003, Israel.
| | - Nicole Werner
- Department of Surgery, University of Wisconsin-Madison School of Medicine and Public Health, 600 Highland Ave, Madison, WI, 53792, USA
| | - Jonathan Kouchly
- Faculty of Data and Decision Sciences, Technion - Israel Institute of Technology, Haifa, 3200003, Israel
| | - Emma Huston
- Clinical Simulation Program, University of Wisconsin Hospitals and Clinics, 600 Highland Ave, Madison, WI, 53792, USA
| | - Shannon DiMarco
- Clinical Simulation Program, University of Wisconsin Hospitals and Clinics, 600 Highland Ave, Madison, WI, 53792, USA
| | - Paul DiMusto
- Department of Surgery, University of Wisconsin-Madison School of Medicine and Public Health, 600 Highland Ave, Madison, WI, 53792, USA
| | - Shlomi Laufer
- Faculty of Data and Decision Sciences, Technion - Israel Institute of Technology, Haifa, 3200003, Israel
| |
Collapse
|
3
|
Goodman ED, Patel KK, Zhang Y, Locke W, Kennedy CJ, Mehrotra R, Ren S, Guan M, Zohar O, Downing M, Chen HW, Clark JZ, Berrigan MT, Brat GA, Yeung-Levy S. Analyzing Surgical Technique in Diverse Open Surgical Videos With Multitask Machine Learning. JAMA Surg 2024; 159:185-192. [PMID: 38055227 PMCID: PMC10701669 DOI: 10.1001/jamasurg.2023.6262] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/23/2023] [Accepted: 09/04/2023] [Indexed: 12/07/2023]
Abstract
Objective To overcome limitations of open surgery artificial intelligence (AI) models by curating the largest collection of annotated videos and to leverage this AI-ready data set to develop a generalizable multitask AI model capable of real-time understanding of clinically significant surgical behaviors in prospectively collected real-world surgical videos. Design, Setting, and Participants The study team programmatically queried open surgery procedures on YouTube and manually annotated selected videos to create the AI-ready data set used to train a multitask AI model for 2 proof-of-concept studies, one generating surgical signatures that define the patterns of a given procedure and the other identifying kinematics of hand motion that correlate with surgeon skill level and experience. The Annotated Videos of Open Surgery (AVOS) data set includes 1997 videos from 23 open-surgical procedure types uploaded to YouTube from 50 countries over the last 15 years. Prospectively recorded surgical videos were collected from a single tertiary care academic medical center. Deidentified videos were recorded of surgeons performing open surgical procedures and analyzed for correlation with surgical training. Exposures The multitask AI model was trained on the AI-ready video data set and then retrospectively applied to the prospectively collected video data set. Main Outcomes and Measures Analysis of open surgical videos in near real-time, performance on AI-ready and prospectively collected videos, and quantification of surgeon skill. Results Using the AI-ready data set, the study team developed a multitask AI model capable of real-time understanding of surgical behaviors-the building blocks of procedural flow and surgeon skill-across space and time. Through principal component analysis, a single compound skill feature was identified, composed of a linear combination of kinematic hand attributes. This feature was a significant discriminator between experienced surgeons and surgical trainees across 101 prospectively collected surgical videos of 14 operators. For each unit increase in the compound feature value, the odds of the operator being an experienced surgeon were 3.6 times higher (95% CI, 1.67-7.62; P = .001). Conclusions and Relevance In this observational study, the AVOS-trained model was applied to analyze prospectively collected open surgical videos and identify kinematic descriptors of surgical skill related to efficiency of hand motion. The ability to provide AI-deduced insights into surgical structure and skill is valuable in optimizing surgical skill acquisition and ultimately improving surgical care.
Collapse
Affiliation(s)
- Emmett D. Goodman
- Department of Computer Science, Stanford University, Stanford, California
- Department of Biomedical Data Science, Stanford University, Stanford, California
| | - Krishna K. Patel
- Department of Computer Science, Stanford University, Stanford, California
- Department of Biomedical Data Science, Stanford University, Stanford, California
| | - Yilun Zhang
- Department of Surgery, Beth Israel Deaconess Medical Center, Boston, Massachusetts
| | - William Locke
- Department of Computer Science, Stanford University, Stanford, California
- Department of Biomedical Data Science, Stanford University, Stanford, California
| | - Chris J. Kennedy
- Department of Surgery, Beth Israel Deaconess Medical Center, Boston, Massachusetts
- Department of Biomedical Informatics, Harvard Medical School, Boston, Massachusetts
| | - Rohan Mehrotra
- Department of Computer Science, Stanford University, Stanford, California
- Department of Biomedical Data Science, Stanford University, Stanford, California
| | - Stephen Ren
- Department of Computer Science, Stanford University, Stanford, California
- Department of Biomedical Data Science, Stanford University, Stanford, California
| | - Melody Guan
- Department of Computer Science, Stanford University, Stanford, California
- Department of Biomedical Data Science, Stanford University, Stanford, California
| | - Orr Zohar
- Department of Biomedical Data Science, Stanford University, Stanford, California
- Department of Electrical Engineering, Stanford University, Stanford, California
| | - Maren Downing
- Department of Surgery, Beth Israel Deaconess Medical Center, Boston, Massachusetts
| | - Hao Wei Chen
- Department of Surgery, Beth Israel Deaconess Medical Center, Boston, Massachusetts
| | - Jevin Z. Clark
- Department of Surgery, Beth Israel Deaconess Medical Center, Boston, Massachusetts
| | - Margaret T. Berrigan
- Department of Surgery, Beth Israel Deaconess Medical Center, Boston, Massachusetts
| | - Gabriel A. Brat
- Department of Surgery, Beth Israel Deaconess Medical Center, Boston, Massachusetts
- Department of Biomedical Informatics, Harvard Medical School, Boston, Massachusetts
| | - Serena Yeung-Levy
- Department of Computer Science, Stanford University, Stanford, California
- Department of Biomedical Data Science, Stanford University, Stanford, California
- Department of Electrical Engineering, Stanford University, Stanford, California
- Clinical Excellence Research Center, Stanford University School of Medicine, Stanford, California
| |
Collapse
|
4
|
Halperin L, Sroka G, Zuckerman I, Laufer S. Automatic performance evaluation of the intracorporeal suture exercise. Int J Comput Assist Radiol Surg 2024; 19:83-86. [PMID: 37278834 DOI: 10.1007/s11548-023-02963-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2023] [Accepted: 05/16/2023] [Indexed: 06/07/2023]
Abstract
PURPOSE This work uses deep learning algorithms to provide automated feedback on the suture with intracorporeal knot exercise in the fundamentals of laparoscopic surgery simulator. Different metrics were designed to provide informative feedback to the user on how to complete the task more efficiently. The automation of the feedback will allow students to practice at any time without the supervision of experts. METHODS Five residents and five senior surgeons participated in the study. Object detection, image classification, and semantic segmentation deep learning algorithms were used to collect statistics on the practitioner's performance. Three task-specific metrics were defined. The metrics refer to the way the practitioner holds the needle before the insertion to the Penrose drain, and the amount of movement of the Penrose drain during the needle's insertion. RESULTS Good agreement between the human labeling and the different algorithms' performance and metric values was achieved. The difference between the scores of the senior surgeons and the surgical residents was statistically significant for one of the metrics. CONCLUSION We developed a system that provides performance metrics of the intracorporeal suture exercise. These metrics can help surgical residents practice independently and receive informative feedback on how they entered the needle into the Penrose.
Collapse
Affiliation(s)
- Liran Halperin
- Faculty of Data and Decision Sciences, Technion - Israel Institute of Technology, 3200003, Haifa, Israel.
| | - Gideon Sroka
- Department of General Surgery, Bnai-Zion Medical Center, Haifa, Israel
| | - Ido Zuckerman
- Faculty of Data and Decision Sciences, Technion - Israel Institute of Technology, 3200003, Haifa, Israel
| | - Shlomi Laufer
- Faculty of Data and Decision Sciences, Technion - Israel Institute of Technology, 3200003, Haifa, Israel
| |
Collapse
|
5
|
Bkheet E, D'Angelo AL, Goldbraikh A, Laufer S. Using hand pose estimation to automate open surgery training feedback. Int J Comput Assist Radiol Surg 2023:10.1007/s11548-023-02947-6. [PMID: 37253925 DOI: 10.1007/s11548-023-02947-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2023] [Accepted: 05/02/2023] [Indexed: 06/01/2023]
Abstract
PURPOSE This research aims to facilitate the use of state-of-the-art computer vision algorithms for the automated training of surgeons and the analysis of surgical footage. By estimating 2D hand poses, we model the movement of the practitioner's hands, and their interaction with surgical instruments, to study their potential benefit for surgical training. METHODS We leverage pre-trained models on a publicly available hands dataset to create our own in-house dataset of 100 open surgery simulation videos with 2D hand poses. We also assess the ability of pose estimations to segment surgical videos into gestures and tool-usage segments and compare them to kinematic sensors and I3D features. Furthermore, we introduce 6 novel surgical dexterity proxies stemming from domain experts' training advice, all of which our framework can automatically detect given raw video footage. RESULTS State-of-the-art gesture segmentation accuracy of 88.35% on the open surgery simulation dataset is achieved with the fusion of 2D poses and I3D features from multiple angles. The introduced surgical skill proxies presented significant differences for novices compared to experts and produced actionable feedback for improvement. CONCLUSION This research demonstrates the benefit of pose estimations for open surgery by analyzing their effectiveness in gesture segmentation and skill assessment. Gesture segmentation using pose estimations achieved comparable results to physical sensors while being remote and markerless. Surgical dexterity proxies that rely on pose estimation proved they can be used to work toward automated training feedback. We hope our findings encourage additional collaboration on novel skill proxies to make surgical training more efficient.
Collapse
Affiliation(s)
- Eddie Bkheet
- Data and Decision Sciences, Technion Institute of Technology, Haifa, Israel.
| | | | - Adam Goldbraikh
- Applied Mathematics, Technion Institute of Technology, Haifa, Israel
| | - Shlomi Laufer
- Data and Decision Sciences, Technion Institute of Technology, Haifa, Israel
| |
Collapse
|
6
|
Video-based formative and summative assessment of surgical tasks using deep learning. Sci Rep 2023; 13:1038. [PMID: 36658186 PMCID: PMC9852463 DOI: 10.1038/s41598-022-26367-9] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2022] [Accepted: 12/13/2022] [Indexed: 01/20/2023] Open
Abstract
To ensure satisfactory clinical outcomes, surgical skill assessment must be objective, time-efficient, and preferentially automated-none of which is currently achievable. Video-based assessment (VBA) is being deployed in intraoperative and simulation settings to evaluate technical skill execution. However, VBA is manual, time-intensive, and prone to subjective interpretation and poor inter-rater reliability. Herein, we propose a deep learning (DL) model that can automatically and objectively provide a high-stakes summative assessment of surgical skill execution based on video feeds and low-stakes formative assessment to guide surgical skill acquisition. Formative assessment is generated using heatmaps of visual features that correlate with surgical performance. Hence, the DL model paves the way for the quantitative and reproducible evaluation of surgical tasks from videos with the potential for broad dissemination in surgical training, certification, and credentialing.
Collapse
|
7
|
Mascagni P, Alapatt D, Sestini L, Altieri MS, Madani A, Watanabe Y, Alseidi A, Redan JA, Alfieri S, Costamagna G, Boškoski I, Padoy N, Hashimoto DA. Computer vision in surgery: from potential to clinical value. NPJ Digit Med 2022; 5:163. [PMID: 36307544 PMCID: PMC9616906 DOI: 10.1038/s41746-022-00707-5] [Citation(s) in RCA: 35] [Impact Index Per Article: 17.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/15/2022] [Accepted: 10/10/2022] [Indexed: 11/09/2022] Open
Abstract
Hundreds of millions of operations are performed worldwide each year, and the rising uptake in minimally invasive surgery has enabled fiber optic cameras and robots to become both important tools to conduct surgery and sensors from which to capture information about surgery. Computer vision (CV), the application of algorithms to analyze and interpret visual data, has become a critical technology through which to study the intraoperative phase of care with the goals of augmenting surgeons' decision-making processes, supporting safer surgery, and expanding access to surgical care. While much work has been performed on potential use cases, there are currently no CV tools widely used for diagnostic or therapeutic applications in surgery. Using laparoscopic cholecystectomy as an example, we reviewed current CV techniques that have been applied to minimally invasive surgery and their clinical applications. Finally, we discuss the challenges and obstacles that remain to be overcome for broader implementation and adoption of CV in surgery.
Collapse
Affiliation(s)
- Pietro Mascagni
- Gemelli Hospital, Catholic University of the Sacred Heart, Rome, Italy.
- IHU-Strasbourg, Institute of Image-Guided Surgery, Strasbourg, France.
- Global Surgical Artificial Intelligence Collaborative, Toronto, ON, Canada.
| | - Deepak Alapatt
- ICube, University of Strasbourg, CNRS, IHU, Strasbourg, France
| | - Luca Sestini
- ICube, University of Strasbourg, CNRS, IHU, Strasbourg, France
- Department of Electronics, Information and Bioengineering, Politecnico di Milano, Milano, Italy
| | - Maria S Altieri
- Global Surgical Artificial Intelligence Collaborative, Toronto, ON, Canada
- Department of Surgery, University of Pennsylvania Perelman School of Medicine, Philadelphia, PA, USA
| | - Amin Madani
- Global Surgical Artificial Intelligence Collaborative, Toronto, ON, Canada
- Department of Surgery, University Health Network, Toronto, ON, Canada
| | - Yusuke Watanabe
- Global Surgical Artificial Intelligence Collaborative, Toronto, ON, Canada
- Department of Surgery, University of Hokkaido, Hokkaido, Japan
| | - Adnan Alseidi
- Global Surgical Artificial Intelligence Collaborative, Toronto, ON, Canada
- Department of Surgery, University of California San Francisco, San Francisco, CA, USA
| | - Jay A Redan
- Department of Surgery, AdventHealth-Celebration Health, Celebration, FL, USA
| | - Sergio Alfieri
- Fondazione Policlinico Universitario A. Gemelli IRCCS, Rome, Italy
| | - Guido Costamagna
- Fondazione Policlinico Universitario A. Gemelli IRCCS, Rome, Italy
| | - Ivo Boškoski
- Fondazione Policlinico Universitario A. Gemelli IRCCS, Rome, Italy
| | - Nicolas Padoy
- IHU-Strasbourg, Institute of Image-Guided Surgery, Strasbourg, France
- ICube, University of Strasbourg, CNRS, IHU, Strasbourg, France
| | - Daniel A Hashimoto
- Global Surgical Artificial Intelligence Collaborative, Toronto, ON, Canada
- Department of Surgery, University of Pennsylvania Perelman School of Medicine, Philadelphia, PA, USA
| |
Collapse
|
8
|
Kil I, Eidt JF, Groff RE, Singapogu RB. Assessment of open surgery suturing skill: Simulator platform, force-based, and motion-based metrics. Front Med (Lausanne) 2022; 9:897219. [PMID: 36111107 PMCID: PMC9468321 DOI: 10.3389/fmed.2022.897219] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2022] [Accepted: 08/05/2022] [Indexed: 11/29/2022] Open
Abstract
Objective This paper focuses on simulator-based assessment of open surgery suturing skill. We introduce a new surgical simulator designed to collect synchronized force, motion, video and touch data during a radial suturing task adapted from the Fundamentals of Vascular Surgery (FVS) skill assessment. The synchronized data is analyzed to extract objective metrics for suturing skill assessment. Methods The simulator has a camera positioned underneath the suturing membrane, enabling visual tracking of the needle during suturing. Needle tracking data enables extraction of meaningful metrics related to both the process and the product of the suturing task. To better simulate surgical conditions, the height of the system and the depth of the membrane are both adjustable. Metrics for assessment of suturing skill based on force/torque, motion, and physical contact are presented. Experimental data are presented from a study comparing attending surgeons and surgery residents. Results Analysis shows force metrics (absolute maximum force/torque in z-direction), motion metrics (yaw, pitch, roll), physical contact metric, and image-enabled force metrics (orthogonal and tangential forces) are found to be statistically significant in differentiating suturing skill between attendings and residents. Conclusion and significance The results suggest that this simulator and accompanying metrics could serve as a useful tool for assessing and teaching open surgery suturing skill.
Collapse
Affiliation(s)
- Irfan Kil
- Department of Electrical and Computer Engineering, Clemson University, Clemson, SC, United States
| | - John F. Eidt
- Division of Vascular Surgery, Baylor Scott & White Heart and Vascular Hospital, Dallas, TX, United States
| | - Richard E. Groff
- Department of Electrical and Computer Engineering, Clemson University, Clemson, SC, United States
| | - Ravikiran B. Singapogu
- Department of Bioengineering, Clemson University, Clemson, SC, United States
- *Correspondence: Ravikiran B. Singapogu
| |
Collapse
|
9
|
Basiev K, Goldbraikh A, Pugh CM, Laufer S. Open surgery tool classification and hand utilization using a multi-camera system. Int J Comput Assist Radiol Surg 2022; 17:1497-1505. [PMID: 35759176 DOI: 10.1007/s11548-022-02691-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2022] [Accepted: 05/26/2022] [Indexed: 11/30/2022]
Abstract
PURPOSE The goal of this work is to use multi-camera video to classify open surgery tools as well as identify which tool is held in each hand. Multi-camera systems help prevent occlusions in open surgery video data. Furthermore, combining multiple views such as a top-view camera covering the full operative field and a close-up camera focusing on hand motion and anatomy may provide a more comprehensive view of the surgical workflow. However, multi-camera data fusion poses a new challenge: A tool may be visible in one camera and not the other. Thus, we defined the global ground truth as the tools being used regardless their visibility. Therefore, tools that are out of the image should be remembered for extensive periods of time while the system responds quickly to changes visible in the video. METHODS Participants (n = 48) performed a simulated open bowel repair. A top-view and a close-up cameras were used. YOLOv5 was used for tool and hand detection. A high-frequency LSTM with a 1-second window at 30 frames per second (fps) and a low-frequency LSTM with a 40-second window at 3 fps were used for spatial, temporal, and multi-camera integration. RESULTS The accuracy and F1 of the six systems were: top-view (0.88/0.88), close-up (0.81,0.83), both cameras (0.9/0.9), high-fps LSTM (0.92/0.93), low-fps LSTM (0.9/0.91), and our final architecture the multi-camera classifier(0.93/0.94). CONCLUSION Since each camera in a multi-camera system may have a partial view of the procedure, we defined a 'global ground truth.' Defining this at the data labeling phase emphasized this requirement at the learning phase, eliminating the need for any heuristic decisions. By combining a system with a high fps and a low fps from the multiple camera array, we improved the classification abilities of the global ground truth.
Collapse
Affiliation(s)
- Kristina Basiev
- Faculty of Industrial Engineering and Management, Technion - Israel Institute of Technology, 3200003, Haifa, Israel.
| | - Adam Goldbraikh
- Applied Mathematics Department, Technion - Israel Institute of Technology, 3200003, Haifa, Israel
| | - Carla M Pugh
- Department of Surgery, Stanford University School of Medicine, Stanford, CA, 94305, USA
| | - Shlomi Laufer
- Faculty of Industrial Engineering and Management, Technion - Israel Institute of Technology, 3200003, Haifa, Israel
| |
Collapse
|
10
|
Goldbraikh A, Volk T, Pugh CM, Laufer S. Using open surgery simulation kinematic data for tool and gesture recognition. Int J Comput Assist Radiol Surg 2022; 17:965-979. [PMID: 35419721 PMCID: PMC10766114 DOI: 10.1007/s11548-022-02615-1] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2021] [Accepted: 03/22/2022] [Indexed: 01/18/2023]
Abstract
PURPOSE The use of motion sensors is emerging as a means for measuring surgical performance. Motion sensors are typically used for calculating performance metrics and assessing skill. The aim of this study was to identify surgical gestures and tools used during an open surgery suturing simulation based on motion sensor data. METHODS Twenty-five participants performed a suturing task on a variable tissue simulator. Electromagnetic motion sensors were used to measure their performance. The current study compares GRU and LSTM networks, which are known to perform well on other kinematic datasets, as well as MS-TCN++, which was developed for video data and was adapted in this work for motion sensors data. Finally, we extended all architectures for multi-tasking. RESULTS In the gesture recognition task the MS-TCN++ has the highest performance with accuracy of [Formula: see text] and F1-Macro of [Formula: see text], edit distance of [Formula: see text] and F1@10 of [Formula: see text] In the tool usage recognition task for the right hand, MS-TCN++ performs the best in most metrics with an accuracy score of [Formula: see text], F1-Macro of [Formula: see text], F1@10 of [Formula: see text], and F1@25 of [Formula: see text]. The multi-task GRU performs best in all metrics in the left-hand case, with an accuracy of [Formula: see text], edit distance of [Formula: see text], F1-Macro of [Formula: see text], F1@10 of [Formula: see text], and F1@25 of [Formula: see text]. CONCLUSION In this study, using motion sensor data, we automatically identified the surgical gestures and the tools used during an open surgery suturing simulation. Our methods may be used for computing more detailed performance metrics and assisting in automatic workflow analysis. MS-TCN++ performed better in gesture recognition as well as right-hand tool recognition, while the multi-task GRU provided better results in the left-hand case. It should be noted that our multi-task GRU network is significantly smaller and has achieved competitive results in the rest of the tasks as well.
Collapse
Affiliation(s)
- Adam Goldbraikh
- Applied Mathematics Department, Technion - Israel Institute of Technology, 3200003, Haifa, Israel.
| | - Tomer Volk
- Faculty of Industrial Engineering and Management, Technion - Israel Institute of Technology, 3200003, Haifa, Israel
| | - Carla M Pugh
- Department of Surgery, Stanford University School of Medicine, Stanford, CA, 610101, USA
| | - Shlomi Laufer
- Faculty of Industrial Engineering and Management, Technion - Israel Institute of Technology, 3200003, Haifa, Israel
| |
Collapse
|