1
|
Byrne CA, Voute LC, Marshall JF. Interobserver agreement during clinical magnetic resonance imaging of the equine foot. Equine Vet J 2024. [PMID: 38946165 DOI: 10.1111/evj.14126] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2023] [Accepted: 06/02/2024] [Indexed: 07/02/2024]
Abstract
BACKGROUND Agreement between experienced observers for assessment of pathology and assessment confidence are poorly documented for magnetic resonance imaging (MRI) of the equine foot. OBJECTIVES To report interobserver agreement for pathology assessment and observer confidence for key anatomical structures of the equine foot during MRI. STUDY DESIGN Exploratory clinical study. METHODS Ten experienced observers (diploma or associate level) assessed 15 equine foot MRI studies acquired from clinical databases of 3 MRI systems. Observers graded pathology in seven key anatomical structures (Grade 1: no pathology, Grade 2: mild pathology, Grade 3: moderate pathology, Grade 4: severe pathology) and provided a grade for their confidence for each pathology assessment (Grade 1: high confidence, Grade 2: moderate confidence, Grade 3: limited confidence, Grade 4: no confidence). Interobserver agreement for the presence/absence of pathology and agreement for individual grades of pathology were assessed with Fleiss' kappa (k). Overall interobserver agreement for pathology was determined using Fleiss' kappa and Kendall's coefficient of concordance (KCC). The distribution of grading was also visualised with bubble charts. RESULTS Interobserver agreement for the presence/absence of pathology of individual anatomical structures was poor-to-fair, except for the navicular bone which had moderate agreement (k = 0.52). Relative agreement for pathology grading (accounting for the ranking of grades) ranged from KCC = 0.19 for the distal interphalangeal joint to KCC = 0.70 for the navicular bone. Agreement was generally greatest at the extremes of pathology. Observer confidence in pathology assessment was generally moderate to high. MAIN LIMITATIONS Distribution of pathology varied between anatomical structures due to random selection of clinical MRI studies. Observers had most experience with low-field MRI. CONCLUSIONS Even with experienced observers, there can be notable variation in the perceived severity of foot pathology on MRI for individual cases, which could be important in a clinical context.
Collapse
Affiliation(s)
- Christian A Byrne
- School of Veterinary Medicine, College of Medical, Veterinary and Life Sciences, University of Glasgow, Glasgow, UK
| | - Lance C Voute
- School of Veterinary Medicine, College of Medical, Veterinary and Life Sciences, University of Glasgow, Glasgow, UK
| | - John F Marshall
- School of Veterinary Medicine, College of Medical, Veterinary and Life Sciences, University of Glasgow, Glasgow, UK
| |
Collapse
|
2
|
Ibragimov B, Mello-Thoms C. The Use of Machine Learning in Eye Tracking Studies in Medical Imaging: A Review. IEEE J Biomed Health Inform 2024; 28:3597-3612. [PMID: 38421842 PMCID: PMC11262011 DOI: 10.1109/jbhi.2024.3371893] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/02/2024]
Abstract
Machine learning (ML) has revolutionized medical image-based diagnostics. In this review, we cover a rapidly emerging field that can be potentially significantly impacted by ML - eye tracking in medical imaging. The review investigates the clinical, algorithmic, and hardware properties of the existing studies. In particular, it evaluates 1) the type of eye-tracking equipment used and how the equipment aligns with study aims; 2) the software required to record and process eye-tracking data, which often requires user interface development, and controller command and voice recording; 3) the ML methodology utilized depending on the anatomy of interest, gaze data representation, and target clinical application. The review concludes with a summary of recommendations for future studies, and confirms that the inclusion of gaze data broadens the ML applicability in Radiology from computer-aided diagnosis (CAD) to gaze-based image annotation, physicians' error detection, fatigue recognition, and other areas of potentially high research and clinical impact.
Collapse
|
3
|
Neves J, Hsieh C, Nobre IB, Sousa SC, Ouyang C, Maciel A, Duchowski A, Jorge J, Moreira C. Shedding light on ai in radiology: A systematic review and taxonomy of eye gaze-driven interpretability in deep learning. Eur J Radiol 2024; 172:111341. [PMID: 38340426 DOI: 10.1016/j.ejrad.2024.111341] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Revised: 01/04/2024] [Accepted: 01/25/2024] [Indexed: 02/12/2024]
Abstract
X-ray imaging plays a crucial role in diagnostic medicine. Yet, a significant portion of the global population lacks access to this essential technology due to a shortage of trained radiologists. Eye-tracking data and deep learning models can enhance X-ray analysis by mapping expert focus areas, guiding automated anomaly detection, optimizing workflow efficiency, and bolstering training methods for novice radiologists. However, the literature shows contradictory results regarding the usefulness of eye-tracking data in deep-learning architectures for abnormality detection. We argue that these discrepancies between studies in the literature are due to (a) the way eye-tracking data is (or is not) processed, (b) the types of deep learning architectures chosen, and (c) the type of application that these architectures will have. We conducted a systematic literature review using PRISMA to address these contradicting results. We analyzed 60 studies that incorporated eye-tracking data in a deep-learning approach for different application goals in radiology. We performed a comparative analysis to understand if eye gaze data contains feature maps that can be useful under a deep learning approach and whether they can promote more interpretable predictions. To the best of our knowledge, this is the first survey in the area that performs a thorough investigation of eye gaze data processing techniques and their impacts in different deep learning architectures for applications such as error detection, classification, object detection, expertise level analysis, fatigue estimation and human attention prediction in medical imaging data. Our analysis resulted in two main contributions: (1) taxonomy that first divides the literature by task, enabling us to analyze the value eye movement can bring for each case and build guidelines regarding architectures and gaze processing techniques adequate for each application, and (2) an overall analysis of how eye gaze data can promote explainability in radiology.
Collapse
Affiliation(s)
- José Neves
- Instituto Superior Técnico / INESC-ID, University of Lisbon, Portugal.
| | - Chihcheng Hsieh
- School of Information Systems, Queensland University of Technology, Australia.
| | | | | | - Chun Ouyang
- School of Information Systems, Queensland University of Technology, Australia.
| | - Anderson Maciel
- Instituto Superior Técnico / INESC-ID, University of Lisbon, Portugal.
| | | | - Joaquim Jorge
- Instituto Superior Técnico / INESC-ID, University of Lisbon, Portugal.
| | - Catarina Moreira
- Human Technology Institute, University of Technology Sydney, Australia.
| |
Collapse
|
4
|
Zang M, Mukund P, Forsyth B, Laine AF, Thakoor KA. Predicting Clinician Fixations on Glaucoma OCT Reports via CNN-Based Saliency Prediction Methods. IEEE OPEN JOURNAL OF ENGINEERING IN MEDICINE AND BIOLOGY 2024; 5:191-197. [PMID: 38606397 PMCID: PMC11008801 DOI: 10.1109/ojemb.2024.3367492] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2024] [Revised: 01/26/2024] [Accepted: 02/15/2024] [Indexed: 04/13/2024] Open
Abstract
Goal: To predict physician fixations specifically on ophthalmology optical coherence tomography (OCT) reports from eye tracking data using CNN based saliency prediction methods in order to aid in the education of ophthalmologists and ophthalmologists-in-training. Methods: Fifteen ophthalmologists were recruited to each examine 20 randomly selected OCT reports and evaluate the likelihood of glaucoma for each report on a scale of 0-100. Eye movements were collected using a Pupil Labs Core eye-tracker. Fixation heat maps were generated using fixation data. Results: A model trained with traditional saliency mapping resulted in a correlation coefficient (CC) value of 0.208, a Normalized Scanpath Saliency (NSS) value of 0.8172, a Kullback-Leibler (KLD) value of 2.573, and a Structural Similarity Index (SSIM) of 0.169. Conclusions: The TranSalNet model was able to predict fixations within certain regions of the OCT report with reasonable accuracy, but more data is needed to improve model accuracy. Future steps include increasing data collection, improving quality of data, and modifying the model architecture.
Collapse
|
5
|
Brunyé TT, Booth K, Hendel D, Kerr KF, Shucard H, Weaver DL, Elmore JG. Machine learning classification of diagnostic accuracy in pathologists interpreting breast biopsies. J Am Med Inform Assoc 2024; 31:552-562. [PMID: 38031453 PMCID: PMC10873842 DOI: 10.1093/jamia/ocad232] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2023] [Revised: 10/19/2023] [Accepted: 11/20/2023] [Indexed: 12/01/2023] Open
Abstract
OBJECTIVE This study explores the feasibility of using machine learning to predict accurate versus inaccurate diagnoses made by pathologists based on their spatiotemporal viewing behavior when evaluating digital breast biopsy images. MATERIALS AND METHODS The study gathered data from 140 pathologists of varying experience levels who each reviewed a set of 14 digital whole slide images of breast biopsy tissue. Pathologists' viewing behavior, including zooming and panning actions, was recorded during image evaluation. A total of 30 features were extracted from the viewing behavior data, and 4 machine learning algorithms were used to build classifiers for predicting diagnostic accuracy. RESULTS The Random Forest classifier demonstrated the best overall performance, achieving a test accuracy of 0.81 and area under the receiver-operator characteristic curve of 0.86. Features related to attention distribution and focus on critical regions of interest were found to be important predictors of diagnostic accuracy. Further including case-level and pathologist-level information incrementally improved classifier performance. DISCUSSION Results suggest that pathologists' viewing behavior during digital image evaluation can be leveraged to predict diagnostic accuracy, affording automated feedback and decision support systems based on viewing behavior to aid in training and, ultimately, clinical practice. They also carry implications for basic research examining the interplay between perception, thought, and action in diagnostic decision-making. CONCLUSION The classifiers developed herein have potential applications in training and clinical settings to provide timely feedback and support to pathologists during diagnostic decision-making. Further research could explore the generalizability of these findings to other medical domains and varied levels of expertise.
Collapse
Affiliation(s)
- Tad T Brunyé
- Center for Applied Brain and Cognitive Sciences, Tufts University, Medford, MA 02155, United States
- Department of Psychology, Tufts University, Medford, MA 02155, United States
| | - Kelsey Booth
- Center for Applied Brain and Cognitive Sciences, Tufts University, Medford, MA 02155, United States
| | - Dalit Hendel
- Center for Applied Brain and Cognitive Sciences, Tufts University, Medford, MA 02155, United States
| | - Kathleen F Kerr
- Department of Biostatistics, University of Washington, Seattle, WA 98105, United States
| | - Hannah Shucard
- Department of Biostatistics, University of Washington, Seattle, WA 98105, United States
| | - Donald L Weaver
- Department of Pathology and Laboratory Medicine, Larner College of Medicine, University of Vermont and Vermont Cancer Center, Burlington, VT 05405, United States
| | - Joann G Elmore
- Department of Medicine, David Geffen School of Medicine, University of California, Los Angeles, Los Angeles, CA 90095, United States
| |
Collapse
|
6
|
Murphy PM. Visual Image Annotation for Bowel Obstruction: Repeatability and Agreement with Manual Annotation and Neural Networks. J Digit Imaging 2023; 36:2179-2193. [PMID: 37278918 PMCID: PMC10502000 DOI: 10.1007/s10278-023-00825-w] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/04/2023] [Revised: 03/21/2023] [Accepted: 03/29/2023] [Indexed: 06/07/2023] Open
Abstract
Bowel obstruction is a common cause of acute abdominal pain. The development of algorithms for automated detection and characterization of bowel obstruction on CT has been limited by the effort required for manual annotation. Visual image annotation with an eye tracking device may mitigate that limitation. The purpose of this study is to assess the agreement between visual and manual annotations for bowel segmentation and diameter measurement, and to assess agreement with convolutional neural networks (CNNs) trained using that data. Sixty CT scans of 50 patients with bowel obstruction from March to June 2022 were retrospectively included and partitioned into training and test data sets. An eye tracking device was used to record 3-dimensional coordinates within the scans, while a radiologist cast their gaze at the centerline of the bowel, and adjusted the size of a superimposed ROI to approximate the diameter of the bowel. For each scan, 59.4 ± 15.1 segments, 847.9 ± 228.1 gaze locations, and 5.8 ± 1.2 m of bowel were recorded. 2d and 3d CNNs were trained using this data to predict bowel segmentation and diameter maps from the CT scans. For comparisons between two repetitions of visual annotation, CNN predictions, and manual annotations, Dice scores for bowel segmentation ranged from 0.69 ± 0.17 to 0.81 ± 0.04 and intraclass correlations [95% CI] for diameter measurement ranged from 0.672 [0.490-0.782] to 0.940 [0.933-0.947]. Thus, visual image annotation is a promising technique for training CNNs to perform bowel segmentation and diameter measurement in CT scans of patients with bowel obstruction.
Collapse
Affiliation(s)
- Paul M Murphy
- University of California-San Diego, 9500 Gilman Dr, 92093, La Jolla, CA, USA.
- UCSD Radiology, 200 W Arbor Dr, 92103, San Diego, CA, USA.
| |
Collapse
|
7
|
Wallerius KP, Bayan SL, Armstrong MF, Lebechi CA, Dey JK, Orbelo DM. Visual Interpretation of Vocal Fold Paralysis in Flexible Laryngoscopy Using Eye Tracking Technology. J Voice 2023:S0892-1997(23)00091-7. [PMID: 37005128 DOI: 10.1016/j.jvoice.2023.02.035] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2022] [Revised: 02/24/2023] [Accepted: 02/24/2023] [Indexed: 04/03/2023]
Abstract
OBJECTIVES Interpretation of laryngoscopy is an important diagnostic skill in otolaryngology. There is, however, limited understanding of the specific visual strategies used while assessing flexible laryngoscopy video. Eye-tracking technology allows for objective study of eye movements during dynamic tasks. The purpose of the present study was to explore visual gaze strategies during laryngoscopy interpretation of unilateral vocal fold paralysis (UVFP) across clinician experience from novice to expert. METHODS Thirty individuals were shown five flexible laryngoscopy videos, each 10 seconds long. After viewing each video, participants reported their impressions of "left vocal fold paralysis," "right vocal fold paralysis," or "no vocal fold paralysis." Eye tracking data were collected and analyzed for duration of fixation and number of fixations on select areas of interest (AOI). Diagnostic accuracy and visual gaze patterns were compared between novice, experienced, and expert groups. RESULTS Diagnostic accuracy among learners in the novice group was significantly lower than those in the more experienced groups (P = 0.04). All groups demonstrated similar visual gaze patterns when viewing the video with normal bilateral vocal fold mobility, spending the greatest percentage of time viewing the trachea. There were differences among groups when viewing the videos of left or right VFP, but the trachea was always in the top three structures for greatest fixation duration and highest number of fixations. CONCLUSIONS Eye-tracking is a novel tool in the setting of laryngoscopy interpretation. With further study it has the potential to be useful for the training of otolaryngology learners to improve diagnostic skills.
Collapse
Affiliation(s)
- Katherine P Wallerius
- Department of Otolaryngology-Head and Neck Surgery, Mayo Clinic, Rochester, Minnesota
| | - Semirra L Bayan
- Department of Otolaryngology-Head and Neck Surgery, Mayo Clinic, Rochester, Minnesota
| | - Michael F Armstrong
- Department of Otolaryngology-Head and Neck Surgery, Mayo Clinic, Rochester, Minnesota
| | - Chiamaka A Lebechi
- Department of Otolaryngology-Head and Neck Surgery, Mayo Clinic, Rochester, Minnesota
| | - Jacob K Dey
- Department of Otolaryngology-Head and Neck Surgery, Mayo Clinic, Rochester, Minnesota
| | - Diana M Orbelo
- Department of Otolaryngology-Head and Neck Surgery, Mayo Clinic, Rochester, Minnesota.
| |
Collapse
|
8
|
Karargyris A, Kashyap S, Lourentzou I, Wu JT, Sharma A, Tong M, Abedin S, Beymer D, Mukherjee V, Krupinski EA, Moradi M. Creation and validation of a chest X-ray dataset with eye-tracking and report dictation for AI development. Sci Data 2021; 8:92. [PMID: 33767191 PMCID: PMC7994908 DOI: 10.1038/s41597-021-00863-5] [Citation(s) in RCA: 19] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2020] [Accepted: 02/09/2021] [Indexed: 12/15/2022] Open
Abstract
We developed a rich dataset of Chest X-Ray (CXR) images to assist investigators in artificial intelligence. The data were collected using an eye-tracking system while a radiologist reviewed and reported on 1,083 CXR images. The dataset contains the following aligned data: CXR image, transcribed radiology report text, radiologist's dictation audio and eye gaze coordinates data. We hope this dataset can contribute to various areas of research particularly towards explainable and multimodal deep learning/machine learning methods. Furthermore, investigators in disease classification and localization, automated radiology report generation, and human-machine interaction can benefit from these data. We report deep learning experiments that utilize the attention maps produced by the eye gaze dataset to show the potential utility of this dataset.
Collapse
Affiliation(s)
| | | | - Ismini Lourentzou
- IBM Research, Almaden Research Center, San Jose, CA, 95120, USA
- Department of Computer Science, Virginia Tech, Blacksburg, VA, 24061, USA
| | - Joy T Wu
- IBM Research, Almaden Research Center, San Jose, CA, 95120, USA
| | - Arjun Sharma
- IBM Research, Almaden Research Center, San Jose, CA, 95120, USA
| | - Matthew Tong
- IBM Research, Almaden Research Center, San Jose, CA, 95120, USA
| | - Shafiq Abedin
- IBM Research, Almaden Research Center, San Jose, CA, 95120, USA
| | - David Beymer
- IBM Research, Almaden Research Center, San Jose, CA, 95120, USA
| | | | - Elizabeth A Krupinski
- Department of Radiology and Imaging Sciences, Emory University, Atlanta, GA, 30322, USA
| | - Mehdi Moradi
- IBM Research, Almaden Research Center, San Jose, CA, 95120, USA.
| |
Collapse
|
9
|
Ramlogan RR, Chuan A, Mariano ER. Contemporary training methods in regional anaesthesia: fundamentals and innovations. Anaesthesia 2021; 76 Suppl 1:53-64. [PMID: 33426656 DOI: 10.1111/anae.15244] [Citation(s) in RCA: 19] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/20/2020] [Indexed: 12/26/2022]
Abstract
Over the past two decades, regional anaesthesia and medical education as a whole have undergone a renaissance. Significant changes in our teaching methods and clinical practice have been influenced by improvements in our theoretical understanding as well as by technological innovations. More recently, there has been a focus on using foundational education principles to teach regional anaesthesia, and the evidence on how to best teach and assess trainees is growing. This narrative review will discuss fundamentals and innovations in regional anaesthesia training. We present the fundamentals in regional anaesthesia training, specifically the current state of simulation-based education, deliberate practice and curriculum design based on competency-based progression. Moving into the future, we present the latest innovations in web-based learning, emerging technologies for teaching and assessment and new developments in alternate reality learning systems.
Collapse
Affiliation(s)
- R R Ramlogan
- Department of Anesthesiology and Pain Medicine, The Ottawa Hospital Research Institute, The Ottawa Hospital, University of Ottawa, Ottawa, Ontario, Canada
| | - A Chuan
- Department of Anaesthesia, Liverpool Hospital, South Western Sydney Clinical School, University of New South Wales, Sydney, New South Wales, Australia
| | - E R Mariano
- Department of Anesthesiology, Peri-operative and Pain Medicine, Stanford University School of Medicine, Stanford, California, USA.,Anesthesiology and Peri-operative Care Service, Veterans Affairs Palo Alto Health Care System, Palo Alto, California, USA
| |
Collapse
|
10
|
Li Y, Cao H, Allen CM, Wang X, Erdelez S, Shyu CR. Computational modeling of human reasoning processes for interpretable visual knowledge: a case study with radiographers. Sci Rep 2020; 10:21620. [PMID: 33303770 PMCID: PMC7730148 DOI: 10.1038/s41598-020-77550-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2020] [Accepted: 11/10/2020] [Indexed: 11/18/2022] Open
Abstract
Visual reasoning is critical in many complex visual tasks in medicine such as radiology or pathology. It is challenging to explicitly explain reasoning processes due to the dynamic nature of real-time human cognition. A deeper understanding of such reasoning processes is necessary for improving diagnostic accuracy and computational tools. Most computational analysis methods for visual attention utilize black-box algorithms which lack explainability and are therefore limited in understanding the visual reasoning processes. In this paper, we propose a computational method to quantify and dissect visual reasoning. The method characterizes spatial and temporal features and identifies common and contrast visual reasoning patterns to extract significant gaze activities. The visual reasoning patterns are explainable and can be compared among different groups to discover strategy differences. Experiments with radiographers of varied levels of expertise on 10 levels of visual tasks were conducted. Our empirical observations show that the method can capture the temporal and spatial features of human visual attention and distinguish expertise level. The extracted patterns are further examined and interpreted to showcase key differences between expertise levels in the visual reasoning processes. By revealing task-related reasoning processes, this method demonstrates potential for explaining human visual understanding.
Collapse
Affiliation(s)
- Yu Li
- Department of Electrical Engineering and Computer Science, University of Missouri, Columbia, MO, 65211, USA
| | - Hongfei Cao
- Department of Electrical Engineering and Computer Science, University of Missouri, Columbia, MO, 65211, USA
| | - Carla M Allen
- Department of Clinical and Diagnostic Science, University of Missouri, Columbia, MO, 65211, USA
| | - Xin Wang
- Department of Information Science, University of Northern Texas, Denton, TX, 76203, USA
| | - Sanda Erdelez
- School of Library and Information Science, Simmons University, Boston, MA, 02115, USA
| | - Chi-Ren Shyu
- Department of Electrical Engineering and Computer Science, University of Missouri, Columbia, MO, 65211, USA.
- Institute for Data Science and Informatics, University of Missouri, Columbia, MO, 65211, USA.
| |
Collapse
|
11
|
Hofmaenner DA, Klinzing S, Brandi G, Hess S, Lohmeyer Q, Enthofer K, Schuepbach RA, Buehler PK. The doctor's point of view: eye-tracking as an investigative tool in the extubation process in intensive care units. A pilot study. Minerva Anestesiol 2020; 86:1180-1189. [DOI: 10.23736/s0375-9393.20.14468-7] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
12
|
Stember JN, Celik H, Krupinski E, Chang PD, Mutasa S, Wood BJ, Lignelli A, Moonis G, Schwartz LH, Jambawalikar S, Bagci U. Eye Tracking for Deep Learning Segmentation Using Convolutional Neural Networks. J Digit Imaging 2020; 32:597-604. [PMID: 31044392 PMCID: PMC6646645 DOI: 10.1007/s10278-019-00220-4] [Citation(s) in RCA: 23] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/12/2022] Open
Abstract
Deep learning with convolutional neural networks (CNNs) has experienced tremendous growth in multiple healthcare applications and has been shown to have high accuracy in semantic segmentation of medical (e.g., radiology and pathology) images. However, a key barrier in the required training of CNNs is obtaining large-scale and precisely annotated imaging data. We sought to address the lack of annotated data with eye tracking technology. As a proof of principle, our hypothesis was that segmentation masks generated with the help of eye tracking (ET) would be very similar to those rendered by hand annotation (HA). Additionally, our goal was to show that a CNN trained on ET masks would be equivalent to one trained on HA masks, the latter being the current standard approach. Step 1: Screen captures of 19 publicly available radiologic images of assorted structures within various modalities were analyzed. ET and HA masks for all regions of interest (ROIs) were generated from these image datasets. Step 2: Utilizing a similar approach, ET and HA masks for 356 publicly available T1-weighted postcontrast meningioma images were generated. Three hundred six of these image + mask pairs were used to train a CNN with U-net-based architecture. The remaining 50 images were used as the independent test set. Step 1: ET and HA masks for the nonneurological images had an average Dice similarity coefficient (DSC) of 0.86 between each other. Step 2: Meningioma ET and HA masks had an average DSC of 0.85 between each other. After separate training using both approaches, the ET approach performed virtually identically to HA on the test set of 50 images. The former had an area under the curve (AUC) of 0.88, while the latter had AUC of 0.87. ET and HA predictions had trimmed mean DSCs compared to the original HA maps of 0.73 and 0.74, respectively. These trimmed DSCs between ET and HA were found to be statistically equivalent with a p value of 0.015. We have demonstrated that ET can create segmentation masks suitable for deep learning semantic segmentation. Future work will integrate ET to produce masks in a faster, more natural manner that distracts less from typical radiology clinical workflow.
Collapse
Affiliation(s)
- J N Stember
- Department of Radiology, Columbia University Medical Center - NYPH, New York, NY, 10032, USA.
| | - H Celik
- The National Institutes of Health, Clinical Center, Bethesda, MD, 20892, USA
| | - E Krupinski
- Department of Radiology & Imaging Sciences, Emory University, Atlanta, GA, 30322, USA
| | - P D Chang
- Department of Radiology, University of California, Irvine, CA, 92697, USA
| | - S Mutasa
- Department of Radiology, Columbia University Medical Center - NYPH, New York, NY, 10032, USA
| | - B J Wood
- The National Institutes of Health, Clinical Center, Bethesda, MD, 20892, USA
| | - A Lignelli
- Department of Radiology, Columbia University Medical Center - NYPH, New York, NY, 10032, USA
| | - G Moonis
- Department of Radiology, Columbia University Medical Center - NYPH, New York, NY, 10032, USA
| | - L H Schwartz
- Department of Radiology, Columbia University Medical Center - NYPH, New York, NY, 10032, USA
| | - S Jambawalikar
- Department of Radiology, Columbia University Medical Center - NYPH, New York, NY, 10032, USA
| | - U Bagci
- Center for Research in Computer Vision, University of Central Florida, 4328 Scorpius St. HEC 221, Orlando, FL, 32816, USA
| |
Collapse
|
13
|
Vaidyanathan P, Prud'hommeaux E, Alm CO, Pelz JB. Computational framework for fusing eye movements and spoken narratives for image annotation. J Vis 2020; 20:13. [PMID: 32678878 PMCID: PMC7424957 DOI: 10.1167/jov.20.7.13] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2019] [Accepted: 10/23/2019] [Indexed: 11/24/2022] Open
Abstract
Despite many recent advances in the field of computer vision, there remains a disconnect between how computers process images and how humans understand them. To begin to bridge this gap, we propose a framework that integrates human-elicited gaze and spoken language to label perceptually important regions in an image. Our work relies on the notion that gaze and spoken narratives can jointly model how humans inspect and analyze images. Using an unsupervised bitext alignment algorithm originally developed for machine translation, we create meaningful mappings between participants' eye movements over an image and their spoken descriptions of that image. The resulting multimodal alignments are then used to annotate image regions with linguistic labels. The accuracy of these labels exceeds that of baseline alignments obtained using purely temporal correspondence between fixations and words. We also find differences in system performances when identifying image regions using clustering methods that rely on gaze information rather than image features. The alignments produced by our framework can be used to create a database of low-level image features and high-level semantic annotations corresponding to perceptually important image regions. The framework can potentially be applied to any multimodal data stream and to any visual domain. To this end, we provide the research community with access to the computational framework.
Collapse
Affiliation(s)
| | | | - Cecilia O. Alm
- College of Liberal Arts, Rochester Institute of Technology, Rochester, NY, USA
| | - Jeff B. Pelz
- Chester F. Carlson Center for Imaging Science, Rochester Institute of Technology, Rochester, NY, USA
| |
Collapse
|
14
|
Khairat S, Coleman C, Ottmar P, Jayachander DI, Bice T, Carson SS. Association of Electronic Health Record Use With Physician Fatigue and Efficiency. JAMA Netw Open 2020; 3:e207385. [PMID: 32515799 PMCID: PMC7284310 DOI: 10.1001/jamanetworkopen.2020.7385] [Citation(s) in RCA: 30] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 12/13/2022] Open
Abstract
IMPORTANCE The use of electronic health records (EHRs) is directly associated with physician burnout. An underlying factor associated with burnout may be EHR-related fatigue owing to insufficient user-centered interface design and suboptimal usability. OBJECTIVE To examine the association between EHR use and fatigue, as measured by pupillometry, and efficiency, as measured by mouse clicks, time, and number of EHR screens, among intensive care unit (ICU) physicians completing a simulation activity in a prominent EHR. DESIGN, SETTING, AND PARTICIPANTS A cross-sectional, simulation-based EHR usability assessment of a leading EHR system was conducted from March 20 to April 5, 2018, among 25 ICU physicians and physician trainees at a southeastern US academic medical center. Participants completed 4 simulation patient cases in the EHR that involved information retrieval and task execution while wearing eye-tracking glasses. Fatigue was quantified through continuous eye pupil data; EHR efficiency was characterized through task completion time, mouse clicks, and EHR screen visits. Data were analyzed from June 1, 2018, to August 31, 2019. MAIN OUTCOMES AND MEASURES Primary outcomes were physician fatigue, measured by pupillometry (with lower scores indicating greater fatigue), and EHR efficiency, measured by task completion times, number of mouse clicks, and number of screens visited during EHR simulation. RESULTS The 25 ICU physicians (13 women; mean [SD] age, 32.1 [6.1] years) who completed a simulation exercise involving 4 patient cases (mean [SD] completion time, 34:43 [11:41] minutes) recorded a total of 14 hours and 27 minutes of EHR activity. All physician participants experienced physiological fatigue at least once during the exercise, and 20 of 25 participants (80%) experienced physiological fatigue within the first 22 minutes of EHR use. Physicians who experienced EHR-related fatigue in 1 patient case were less efficient in the subsequent patient case, as demonstrated by longer task completion times (r = -0.521; P = .007), higher numbers of mouse clicks (r = -0.562; P = .003), and more EHR screen visits (r = -0.486; P = .01). CONCLUSIONS AND RELEVANCE This study reports high rates of fatigue among ICU physicians during short periods of EHR simulation, which were negatively associated with EHR efficiency and included a carryover association across patient cases. More research is needed to investigate the underlying causes of EHR-associated fatigue, to support user-centered EHR design, and to inform safe EHR use policies and guidelines.
Collapse
Affiliation(s)
- Saif Khairat
- Carolina Health Informatics Program, University of North Carolina at Chapel Hill
- School of Nursing, University of North Carolina at Chapel Hill
| | - Cameron Coleman
- Department of Preventive Medicine, University of North Carolina at Chapel Hill
| | - Paige Ottmar
- Gilling’s School of Public Health, University of North Carolina at Chapel Hill
| | | | - Thomas Bice
- Pulmonary Diseases and Critical Care Medicine, University of North Carolina at Chapel Hill
| | - Shannon S. Carson
- Pulmonary Diseases and Critical Care Medicine, University of North Carolina at Chapel Hill
| |
Collapse
|
15
|
Castro D, Yang J, Patel P, Sauerbrei E, Hopman W, Kolar M, Soboleski D. Factors affecting perception of the normal pediatric appendix on sonography. Ultrasound J 2019; 11:33. [PMID: 31865464 PMCID: PMC6925608 DOI: 10.1186/s13089-019-0148-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2019] [Accepted: 12/12/2019] [Indexed: 11/17/2022] Open
Abstract
Background To determine if an inherent perception skill along with sonographer experience, knowledge base, scanning time play a role in the identification of the normal appendix in the pediatric population. This is a retrospective review of pediatric (< 18 years old) patients with a clinical suspicion of acute appendicitis presenting to the emergency department of two affiliated academic tertiary care hospitals over a 1-year time span. All patients had a sonogram performed by 1/15 sonographers or by 1/8 on-call radiology residents. Those with a normal or non-visualized appendix with subsequent discharge from ER were included in the study. Patient demographics, minutes spent scanning, and sonographer years of experience in general abdominal scanning and residents level of training were recorded. Results Of the 127 patients included in the study, 51 (40%) were male and 76 (60%) were female, with a mean age of 11.8 ± 4.2 years. Sonographers who failed to see a normal appendix had less experience (median 8 years) than those who did visualize the appendix (median 15 years), p ≤ 0.001. Longer time spent scanning was also associated with visualizing a normal appendix (20.4 versus 29.1 min, p = 0.001). In multivariable logistic regression, more time spent scanning (OR 1.04, 95% CI 1.01, 1.07, p = 0.012) and increased sonographer experience (OR 1.07, 95% CI 1.02, 1.13, p = 0.012) resulted in greater odds of perceiving the appendix. The top 4 were significantly more likely to visualize the appendix (88.0%) than all of the other combined (20.8%, p < 0.001), and they also had substantially more experience (median 15 years versus 8 years, p < 0.001). Overall, sonographers were more likely to see a normal appendix (61%) than the residents (14%), p < 0.001. Conclusion Sonography to rule out appendicitis in the pediatric patient is in general most successful when performed by experienced sonographers with adequate time to perform the scan. Triaging patients to those sonographers who have displayed optimal perceptual ability of the normal appendix may help optimize patient care and hospital resources. Having experienced sonographers available after hours would allow for optimal care in the setting of ‘query’ appendicitis.
Collapse
Affiliation(s)
- Denise Castro
- Department of Radiology, Kingston Health Sciences Centre, Queen's University, 76 Stuart Street, Kingston, ON, K7L 2V7, Canada
| | - Joseph Yang
- Queen's School of Medicine, Queen's University, 80 Barrie Street, Kingston, ON, K7L 3N6, Canada.
| | - Prasan Patel
- Department of Radiology, Kingston Health Sciences Centre, Queen's University, 76 Stuart Street, Kingston, ON, K7L 2V7, Canada
| | - Eric Sauerbrei
- Department of Radiology, Kingston Health Sciences Centre, Queen's University, 76 Stuart Street, Kingston, ON, K7L 2V7, Canada
| | - Wilma Hopman
- WJ Henderson Centre for Patient Oriented Research, Queen's University, 76 Stuart Street, Kingston, ON, K7L 2V7, Canada
| | - Mila Kolar
- Department of Surgery, Kingston Health Sciences Centre, Queen's University, 76 Stuart Street, Kingston, ON, K7L 2V7, Canada
| | - Don Soboleski
- Department of Radiology, Kingston Health Sciences Centre, Queen's University, 76 Stuart Street, Kingston, ON, K7L 2V7, Canada
| |
Collapse
|
16
|
Analysis of the Visual Perception of Female Breast Aesthetics and Symmetry. Plast Reconstr Surg 2019; 144:1257-1266. [DOI: 10.1097/prs.0000000000006292] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
17
|
Grimm LJ, Shelby RA, Knippa EE, Langman EL, Miller LS, Whiteside BA, Soo MS. Frequency of Breast Cancer Thoughts and Lifetime Risk Estimates: A Multi-Institutional Survey of Women Undergoing Screening Mammography. J Am Coll Radiol 2019; 16:1393-1400. [DOI: 10.1016/j.jacr.2018.12.031] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2018] [Revised: 12/11/2018] [Accepted: 12/19/2018] [Indexed: 11/30/2022]
|
18
|
Davies A, Harper S, Vigo M, Jay C. Investigating the effect of clinical history before electrocardiogram interpretation on the visual behavior and interpretation accuracy of clinicians. Sci Rep 2019; 9:11300. [PMID: 31383896 PMCID: PMC6683299 DOI: 10.1038/s41598-019-47830-0] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2018] [Accepted: 07/04/2019] [Indexed: 11/09/2022] Open
Abstract
We examine the impact of the presentation of a patient's clinical history on subsequent visual appraisal and interpretation accuracy of electrocardiograms (ECGs). Healthcare-practitioners (N = 31) skilled in 12-lead ECG interpretation took part in a repeated-measures experiment with counterbalancing viewing 9 ECGs on a computer screen in two separate conditions: with/without an associated patient-history. A Hellinger-distance calculation was applied using a permutation test to eye-movement transitions at two granularity levels: between the ECG leads, and between smaller grid-cells, whose size was determined via data-driven clustering of the fixation points. Findings indicate that presentation of clinical-history does affect accuracy of interpretation in one ECG. Visual-behavior differed as a function of both history presentation and accuracy when considering transitions between the data-driven grid units (using a fine granularity, and able to show attention to parts of the waveform). Differences in visual-behavior at waveform level demonstrate an influence of patient-history and expertise that are not detected at the lead level. Visual-behaviour differs according to whether a patient-history is presented, and whether a clinician provides an accurate interpretation. This difference is evident in how the waveform itself is viewed, and is less present at the coarse granularity of visual transitions between leads. To understand how clinicians interpret ECGs, and potentially other medical images, visual transitions should be considered at a fine level of granularity, determined in a data-driven fashion.
Collapse
Affiliation(s)
- Alan Davies
- School of Computer Science, University of Manchester, Manchester, UK.
| | - Simon Harper
- School of Computer Science, University of Manchester, Manchester, UK
| | - Markel Vigo
- School of Computer Science, University of Manchester, Manchester, UK
| | - Caroline Jay
- School of Computer Science, University of Manchester, Manchester, UK
| |
Collapse
|
19
|
Abstract
Breast cancer is the most common cancer among females worldwide and large volumes of breast images are produced and interpreted annually. As long as radiologists interpret these images, the diagnostic accuracy will be limited by human factors and both false-positive and false-negative errors might occur. By understanding visual search in breast images, we may be able to identify causes of diagnostic errors, find ways to reduce them, and also provide a better education to radiology residents. Many visual search studies in breast radiology have been devoted to mammography. These studies showed that 70% of missed lesions on mammograms attract radiologists' visual attention and that a plethora of different reasons, such as satisfaction of search, incorrect background sampling, and incorrect first impression can cause diagnostic errors in the interpretation of mammograms. Recently, highly accurate tools, which rely on both eye-tracking data and the content of the mammogram, have been proposed to provide feedback to the radiologists. Improving these tools and determining the optimal pathway to integrate them in the radiology workflow could be a possible line of future research. Moreover, in the past few years deep learning has led to improving diagnostic accuracy of computerized diagnostic tools and visual search studies will be required to understand how radiologists interact with the prompts from these tools, and to identify the best way to utilize them. Visual search in other breast imaging modalities, such as breast ultrasound and digital breast tomosynthesis, have so far received less attention, probably due to associated complexities of eye-tracking monitoring and analysing the data. For example, in digital breast tomosynthesis, scrolling through the image results in longer trials, adds a new factor to the study's complexity and makes calculation of gaze parameters more difficult. However, considering the wide utilization of three-dimensional imaging modalities, more visual search studies involving reading stack-view examinations are required in the future. To conclude, in the past few decades visual search studies provided extensive understanding about underlying reasons for diagnostic errors in breast radiology and characterized differences between experts' and novices' visual search patterns. Further visual search studies are required to investigate radiologists' interaction with relatively newer imaging modalities and artificial intelligence tools.
Collapse
Affiliation(s)
- Ziba Gandomkar
- BreastScreen Reader Assessment Strategy (BREAST), Discipline of Medical Imaging Sciences, Faculty of Medicine and Health, University of Sydney, Sydney, NSW, Australia
| | - Claudia Mello-Thoms
- Department of Radiology, Carver College of Medicine, University of Iowa, Iowa City, IA, US
| |
Collapse
|
20
|
Mercan E, Shapiro LG, Brunyé TT, Weaver DL, Elmore JG. Characterizing Diagnostic Search Patterns in Digital Breast Pathology: Scanners and Drillers. J Digit Imaging 2019; 31:32-41. [PMID: 28681097 DOI: 10.1007/s10278-017-9990-5] [Citation(s) in RCA: 16] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022] Open
Abstract
Following a baseline demographic survey, 87 pathologists interpreted 240 digital whole slide images of breast biopsy specimens representing a range of diagnostic categories from benign to atypia, ductal carcinoma in situ, and invasive cancer. A web-based viewer recorded pathologists' behaviors while interpreting a subset of 60 randomly selected and randomly ordered slides. To characterize diagnostic search patterns, we used the viewport location, time stamp, and zoom level data to calculate four variables: average zoom level, maximum zoom level, zoom level variance, and scanning percentage. Two distinct search strategies were confirmed: scanning is characterized by panning at a constant zoom level, while drilling involves zooming in and out at various locations. Statistical analysis was applied to examine the associations of different visual interpretive strategies with pathologist characteristics, diagnostic accuracy, and efficiency. We found that females scanned more than males, and age was positively correlated with scanning percentage, while the facility size was negatively correlated. Throughout 60 cases, the scanning percentage and total interpretation time per slide decreased, and these two variables were positively correlated. The scanning percentage was not predictive of diagnostic accuracy. Increasing average zoom level, maximum zoom level, and zoom variance were correlated with over-interpretation.
Collapse
Affiliation(s)
- Ezgi Mercan
- Paul G. Allen School of Computer Science and Engineering, University of Washington, Seattle, WA, USA.
| | - Linda G Shapiro
- Paul G. Allen School of Computer Science and Engineering, University of Washington, Seattle, WA, USA
| | - Tad T Brunyé
- Department of Psychology, Tufts University, Medford, MA, USA
| | - Donald L Weaver
- Department of Pathology and UVM Cancer Center, University of Vermont, Burlington, VT, USA
| | - Joann G Elmore
- Department of Medicine, University of Washington School of Medicine, Seattle, WA, USA
| |
Collapse
|
21
|
Khosravan N, Celik H, Turkbey B, Jones EC, Wood B, Bagci U. A collaborative computer aided diagnosis (C-CAD) system with eye-tracking, sparse attentional model, and deep learning. Med Image Anal 2019; 51:101-115. [PMID: 30399507 PMCID: PMC6407631 DOI: 10.1016/j.media.2018.10.010] [Citation(s) in RCA: 33] [Impact Index Per Article: 6.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2017] [Revised: 07/27/2018] [Accepted: 10/26/2018] [Indexed: 12/19/2022]
Abstract
Computer aided diagnosis (CAD) tools help radiologists to reduce diagnostic errors such as missing tumors and misdiagnosis. Vision researchers have been analyzing behaviors of radiologists during screening to understand how and why they miss tumors or misdiagnose. In this regard, eye-trackers have been instrumental in understanding visual search processes of radiologists. However, most relevant studies in this aspect are not compatible with realistic radiology reading rooms. In this study, we aim to develop a paradigm shifting CAD system, called collaborative CAD (C-CAD), that unifies CAD and eye-tracking systems in realistic radiology room settings. We first developed an eye-tracking interface providing radiologists with a real radiology reading room experience. Second, we propose a novel algorithm that unifies eye-tracking data and a CAD system. Specifically, we present a new graph based clustering and sparsification algorithm to transform eye-tracking data (gaze) into a graph model to interpret gaze patterns quantitatively and qualitatively. The proposed C-CAD collaborates with radiologists via eye-tracking technology and helps them to improve their diagnostic decisions. The C-CAD uses radiologists' search efficiency by processing their gaze patterns. Furthermore, the C-CAD incorporates a deep learning algorithm in a newly designed multi-task learning platform to segment and diagnose suspicious areas simultaneously. The proposed C-CAD system has been tested in a lung cancer screening experiment with multiple radiologists, reading low dose chest CTs. Promising results support the efficiency, accuracy and applicability of the proposed C-CAD system in a real radiology room setting. We have also shown that our framework is generalizable to more complex applications such as prostate cancer screening with multi-parametric magnetic resonance imaging (mp-MRI).
Collapse
Affiliation(s)
- Naji Khosravan
- Center for Research in Computer Vision, University of Central Florida, FL, United States
| | - Haydar Celik
- Clinical Center, National Institutes of Health, Bethesda, MD, United States
| | - Baris Turkbey
- Clinical Center, National Institutes of Health, Bethesda, MD, United States
| | - Elizabeth C Jones
- Clinical Center, National Institutes of Health, Bethesda, MD, United States
| | - Bradford Wood
- Clinical Center, National Institutes of Health, Bethesda, MD, United States
| | - Ulas Bagci
- Center for Research in Computer Vision, University of Central Florida, FL, United States.
| |
Collapse
|
22
|
Alamudun F, Paulus P, Yoon HJ, Tourassi G. Modeling sequential context effects in diagnostic interpretation of screening mammograms. J Med Imaging (Bellingham) 2018; 5:031408. [PMID: 29564370 PMCID: PMC5858736 DOI: 10.1117/1.jmi.5.3.031408] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2017] [Accepted: 02/19/2018] [Indexed: 11/29/2022] Open
Abstract
Prior research has shown that physicians’ medical decisions can be influenced by sequential context, particularly in cases where successive stimuli exhibit similar characteristics when analyzing medical images. This type of systematic error is known to psychophysicists as sequential context effect as it indicates that judgments are influenced by features of and decisions about the preceding case in the sequence of examined cases, rather than being based solely on the peculiarities unique to the present case. We determine if radiologists experience some form of context bias, using screening mammography as the use case. To this end, we explore correlations between previous perceptual behavior and diagnostic decisions and current decisions. We hypothesize that a radiologist’s visual search pattern and diagnostic decisions in previous cases are predictive of the radiologist’s current diagnostic decisions. To test our hypothesis, we tasked 10 radiologists of varied experience to conduct blind reviews of 100 four-view screening mammograms. Eye-tracking data and diagnostic decisions were collected from each radiologist under conditions mimicking clinical practice. Perceptual behavior was quantified using the fractal dimension of gaze scanpath, which was computed using the Minkowski–Bouligand box-counting method. To test the effect of previous behavior and decisions, we conducted a multifactor fixed-effects ANOVA. Further, to examine the predictive value of previous perceptual behavior and decisions, we trained and evaluated a predictive model for radiologists’ current diagnostic decisions. ANOVA tests showed that previous visual behavior, characterized by fractal analysis, previous diagnostic decisions, and image characteristics of previous cases are significant predictors of current diagnostic decisions. Additionally, predictive modeling of diagnostic decisions showed an overall improvement in prediction error when the model is trained on additional information about previous perceptual behavior and diagnostic decisions.
Collapse
Affiliation(s)
- Folami Alamudun
- Oak Ridge National Laboratory, Computational Sciences and Engineering Division, Oak Ridge, Tennessee, United States.,Oak Ridge National Laboratory, Health Data Sciences Institute, Oak Ridge, Tennessee, United States
| | - Paige Paulus
- University of Tennessee, Department of Mechanical, Aerospace, and Biomedical Engineering, Knoxville, Tennessee, United States
| | - Hong-Jun Yoon
- Oak Ridge National Laboratory, Computational Sciences and Engineering Division, Oak Ridge, Tennessee, United States.,Oak Ridge National Laboratory, Health Data Sciences Institute, Oak Ridge, Tennessee, United States
| | - Georgia Tourassi
- Oak Ridge National Laboratory, Computational Sciences and Engineering Division, Oak Ridge, Tennessee, United States.,Oak Ridge National Laboratory, Health Data Sciences Institute, Oak Ridge, Tennessee, United States.,University of Tennessee, Department of Mechanical, Aerospace, and Biomedical Engineering, Knoxville, Tennessee, United States
| |
Collapse
|
23
|
Ashraf H, Sodergren MH, Merali N, Mylonas G, Singh H, Darzi A. Eye-tracking technology in medical education: A systematic review. MEDICAL TEACHER 2018; 40:62-69. [PMID: 29172823 DOI: 10.1080/0142159x.2017.1391373] [Citation(s) in RCA: 69] [Impact Index Per Article: 11.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/24/2023]
Abstract
BACKGROUND Eye-tracking technology is an established research tool within allied industries such as advertising, psychology and aerospace. This review aims to consolidate literature describing the evidence for use of eye-tracking as an adjunct to traditional teaching methods in medical education. METHODS A systematic literature review was conducted in line with STORIES guidelines. A search of EMBASE, OVID MEDLINE, PsycINFO, TRIP database, and Science Direct was conducted until January 2017. Studies describing the use of eye-tracking in the training, assessment, and feedback of clinicians were included in the review. RESULTS Thirty-three studies were included in the final qualitative synthesis. Three studies were based on the use of gaze training, three studies on the changes in gaze behavior during the learning curve, 17 studies on clinical assessment and six studies focused on the use of eye-tracking methodology as a feedback tool. The studies demonstrated feasibility and validity in the use of eye-tracking as a training and assessment method. CONCLUSIONS Overall, eye-tracking methodology has contributed significantly to the training, assessment, and feedback practices used in the clinical setting. The technology provides reliable quantitative data, which can be interpreted to give an indication of clinical skill, provide training solutions and aid in feedback and reflection. This review provides a detailed summary of evidence relating to eye-tracking methodology and its uses as a training method, changes in visual gaze behavior during the learning curve, eye-tracking methodology for proficiency assessment and its uses as a feedback tool.
Collapse
Affiliation(s)
- Hajra Ashraf
- a Department of Surgery and Cancer , Imperial College, St Mary's Hospital , London , UK
| | - Mikael H Sodergren
- a Department of Surgery and Cancer , Imperial College, St Mary's Hospital , London , UK
| | | | - George Mylonas
- a Department of Surgery and Cancer , Imperial College, St Mary's Hospital , London , UK
| | - Harsimrat Singh
- a Department of Surgery and Cancer , Imperial College, St Mary's Hospital , London , UK
| | - Ara Darzi
- a Department of Surgery and Cancer , Imperial College, St Mary's Hospital , London , UK
| |
Collapse
|
24
|
Bhavsar P, Srinivasan B, Srinivasan R. Quantifying situation awareness of control room operators using eye-gaze behavior. Comput Chem Eng 2017. [DOI: 10.1016/j.compchemeng.2017.06.004] [Citation(s) in RCA: 29] [Impact Index Per Article: 4.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
25
|
Ebner L, Tall M, Choudhury KR, Ly DL, Roos JE, Napel S, Rubin GD. Variations in the functional visual field for detection of lung nodules on chest computed tomography: Impact of nodule size, distance, and local lung complexity. Med Phys 2017; 44:3483-3490. [PMID: 28419484 DOI: 10.1002/mp.12277] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2016] [Revised: 02/24/2017] [Accepted: 03/20/2017] [Indexed: 11/11/2022] Open
Abstract
PURPOSE To explore the characteristics that impact lung nodule detection by peripheral vision when searching for lung nodules on chest CT-scans. METHODS This study was approved by the local IRB and is HIPAA compliant. A simulated primary (1°) target mass (2 × 2 × 5 cm) was embedded into 5 cm thick subvolumes (SV) extracted from three unenhanced lung MDCT scans (64 row, 1.25 mm thickness, 0.7 mm increment). One of 30 solid, secondary nodules with either 3-4 mm and 5-8 mm diameters were embedded into 192 of 207 SVs. The secondary nodule was placed at a random depth within each SV, a transverse distance of 2.5, 5, 7.5, or 10 mm, and along one of eight rays cast every 45° from the center of the 1° mass. Video recordings of transverse paging in cranio-caudal direction were created for each SV (frame rate three sections/sec). Six radiologists observed each cine-loop once while gaze-tracking hardware assured that gaze was centered on the 1° mass. Each radiologist assigned a confidence rating (0-5) to the detection of a secondary nodule and indicated its location. Detection sensitivity was analyzed relative to secondary nodule size, transverse distance, radial orientation, and lung complexity. Lung complexity was characterized by the number of particles (connected pixels) and the sum of the area of all particles above a -500 HU threshold within regions of interest around the 1° mass and secondary nodule. RESULTS Using a proportional odds logistic regression model and eliminating redundant predictors, models fit individually to each reader resulted in the following decreasing order of association based on greatest reduction in Akaike Information Criterion: secondary nodule diameter (6/6 readers, P < 0.001), distance from central mass (6/6 readers, P < 0.001), lung complexity particle count (5/6 readers, P = 0.05), and lung complexity particle area (3/6 readers, P = 0.03). Substantial inter-reader differences in sensitivity to decreasing nodule diameter, distance, and complexity characteristics were observed. CONCLUSIONS Of the investigated parameters, secondary nodule size, distance from the gaze center and lung complexity (particle number and area) significantly impact nodule detection with peripheral vision.
Collapse
Affiliation(s)
- Lukas Ebner
- Department of Radiology, Duke University Medical Center, Durham, NC, 27710, USA
| | - Martin Tall
- Department of Radiology, Duke University Medical Center, Durham, NC, 27710, USA
| | | | - Donald L Ly
- Department of Radiology, Stanford School of Medicine, Stanford, CA, 94305, USA
| | - Justus E Roos
- Department of Radiology, Duke University Medical Center, Durham, NC, 27710, USA
| | - Sandy Napel
- Department of Radiology, Stanford School of Medicine, Stanford, CA, 94305, USA
| | - Geoffrey D Rubin
- Department of Radiology, Duke University Medical Center, Durham, NC, 27710, USA
| |
Collapse
|
26
|
Gandomkar Z, Tay K, Ryder W, Brennan PC, Mello-Thoms C. iCAP: An Individualized Model Combining Gaze Parameters and Image-Based Features to Predict Radiologists' Decisions While Reading Mammograms. IEEE TRANSACTIONS ON MEDICAL IMAGING 2017; 36:1066-1075. [PMID: 28055858 DOI: 10.1109/tmi.2016.2645881] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
This study introduces an individualized tool for identifying mammogram interpretation errors, called eye-Computer Assisted Perception (iCAP). iCAP consists of two modules, one which processes areas marked by radiologists as suspicious for cancer and classifies these as False Positive (FP) or True Positive (TP) decisions, while the second module classifies fixated but not marked locations as False Negative (FN) or True-Negative (TN) decisions. iCAP relies on both radiologists' gaze-related parameters, extracted from eye tracking data, and image-based features. In order to evaluate iCAP, eye tracking data from eight breast radiologists reading 120 two-view digital mammograms were collected. Fifty-nine cases had biopsy proven cancer. For each radiologist, a user-specific support vector machine model was built to classify the radiologist' s reported areas as TPs or FPs and fixated locations as TNs or FNs. The performances of the classifiers were evaluated by utilizing leave-one-out cross validation. iCAP was tested retrospectively in a simulated scenario in which it was assumed that the radiologists would accept all iCAP decisions. Using iCAP led to an average increase of 12%±6% in the number of correctly localized cancer and an average decrease of 44.5%±22.7% in the number of FPs per image.
Collapse
|
27
|
|
28
|
Alamudun F, Yoon HJ, Hudson KB, Morin-Ducote G, Hammond T, Tourassi GD. Fractal analysis of visual search activity for mass detection during mammographic screening. Med Phys 2017; 44:832-846. [PMID: 28079249 DOI: 10.1002/mp.12100] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/24/2016] [Revised: 10/18/2016] [Accepted: 12/20/2016] [Indexed: 11/10/2022] Open
Abstract
PURPOSE The objective of this study was to assess the complexity of human visual search activity during mammographic screening using fractal analysis and to investigate its relationship with case and reader characteristics. METHODS The study was performed for the task of mammographic screening with simultaneous viewing of four coordinated breast views as typically done in clinical practice. Eye-tracking data and diagnostic decisions collected for 100 mammographic cases (25 normal, 25 benign, 50 malignant) from 10 readers (three board certified radiologists and seven Radiology residents), formed the corpus for this study. The fractal dimension of the readers' visual scanning pattern was computed with the Minkowski-Bouligand box-counting method and used as a measure of gaze complexity. Individual factor and group-based interaction ANOVA analysis was performed to study the association between fractal dimension, case pathology, breast density, and reader experience level. The consistency of the observed trends depending on gaze data representation was also examined. RESULTS Case pathology, breast density, reader experience level, and individual reader differences are all independent predictors of the complexity of visual scanning pattern when screening for breast cancer. No higher order effects were found to be significant. CONCLUSIONS Fractal characterization of visual search behavior during mammographic screening is dependent on case properties and image reader characteristics.
Collapse
Affiliation(s)
- Folami Alamudun
- Biomedical Sciences, Engineering, and Computing Group, Health Data Sciences Institute, Oak Ridge National Laboratory, Oak Ridge, TN, 37831, USA
| | - Hong-Jun Yoon
- Biomedical Sciences, Engineering, and Computing Group, Health Data Sciences Institute, Oak Ridge National Laboratory, Oak Ridge, TN, 37831, USA
| | - Kathleen B Hudson
- Department of Radiology, University of Tennessee Medical Center at Knoxville, Knoxville, TN, 37920, USA
| | - Garnetta Morin-Ducote
- Department of Radiology, University of Tennessee Medical Center at Knoxville, Knoxville, TN, 37920, USA
| | - Tracy Hammond
- Department of Computer Science and Engineering, Texas A&M University, College Station, TX, USA, 77843
| | - Georgia D Tourassi
- Biomedical Sciences, Engineering, and Computing Group, Health Data Sciences Institute, Oak Ridge National Laboratory, Oak Ridge, TN, 37831, USA
| |
Collapse
|
29
|
Hanley J, Warren D, Glass N, Tranel D, Karam M, Buckwalter J. Visual Interpretation of Plain Radiographs in Orthopaedics Using Eye-Tracking Technology. THE IOWA ORTHOPAEDIC JOURNAL 2017; 37:225-231. [PMID: 28852362 PMCID: PMC5508291] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
BACKGROUND Despite the importance of radiographic interpretation in orthopaedics, there not a clear understanding of the specific visual strategies used while analyzing a plain film. Eyetracking technology allows for the objective study of eye movements while performing a dynamic task, such as reading X-rays. Our study looks to elucidate objective differences in image interpretation between novice and experienced orthopaedic trainees using this novel technology. METHODS Novice and experienced orthopaedic trainees (N=23) were asked to interpret AP pelvis films, searching for unilateral acetabular fractures while eye-movements were assessed for pattern of gaze, fixation on regions of interest, and time of fixation at regions of interest. Participants were asked to label radiographs as "fractured" or "not fractured." If "fractured", the participant was asked to determine the fracture pattern. A control condition employed Ekman faces and participants judged gender and facial emotion. Data were analyzed for variation in eye movements between participants, accuracy of responses, and response time. RESULTS Accuracy: There was no significant difference by level of training for accurately identifing fracture images (p=0.3255). There was a significant association between higher level of training and correctly identifying non-fractured images (p=0.0155); greater training was also associated with more success in identifying the correct Judet-Letournel classification (p=0.0029). Response Time: Greater training was associated with faster response times (p=0.0009 for fracture images and 0.0012 for non-fractured images). Fixation Duration: There was no correlation of average fixation duration with experience (p=0.9632). Regions of Interest (ROIs): More experience was associated with an average of two fewer fixated ROIs (p=0.0047). Number of Fixations: Increased experience was associated with fewer fixations overall (p=0.0007). CONCLUSIONS Experience has a significant impact on both accuracy and efficiency in interpreting plain films. Greater training is associated with a shift toward a more efficient and thorough assessment of plain radiographs. Eyetracking is a useful descriptive tool in the setting of plain film interpretation. CLINICAL RELEVANCE We propose further assessment of eye movements in larger populations of orthopaedic surgeons, including staff orthopaedists. Describing the differences between novice and expert interpretation may provide insight into ways to accelerate the learning process in young orthopaedists.
Collapse
Affiliation(s)
- Jessica Hanley
- University of Iowa, Department of Orthopaedics and Rehabilitation, Iowa City, IA
| | - David Warren
- University of Nebraska, Department of Neurological Sciences,Omaha, NE
| | - Natalie Glass
- University of Iowa, Department of Orthopaedics and Rehabilitation, Iowa City, IA
| | - Daniel Tranel
- University of Iowa, Department of Neurology, Psychological & Brain Sciences,Iowa City, IA
| | - Matthew Karam
- University of Iowa, Department of Orthopaedics and Rehabilitation, Iowa City, IA
| | - Joseph Buckwalter
- University of Iowa, Department of Orthopaedics and Rehabilitation, Iowa City, IA
| |
Collapse
|
30
|
Waite S, Kolla S, Jeudy J, Legasto A, Macknik SL, Martinez-Conde S, Krupinski EA, Reede DL. Tired in the Reading Room: The Influence of Fatigue in Radiology. J Am Coll Radiol 2016; 14:191-197. [PMID: 27956140 DOI: 10.1016/j.jacr.2016.10.009] [Citation(s) in RCA: 53] [Impact Index Per Article: 6.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/01/2023]
Abstract
Commonly conflated with sleepiness, fatigue is a distinct multidimensional condition with physical and mental effects. Fatigue in health care providers and any secondary effects on patient care are an important societal concern. As medical image interpretation is highly dependent on visual input, visual fatigue is of particular interest to radiologists. Humans analyze their surroundings with rapid eye movements called saccades, and fatigue decreases saccadic velocity. Oculomotor parameters may, therefore, be an objective and reproducible metric of fatigue and eye movement analysis can provide valuable insight into the etiology of fatigue-related error.
Collapse
Affiliation(s)
- Stephen Waite
- Department of Radiology, SUNY Downstate Medical Center, Brooklyn, New York.
| | - Srinivas Kolla
- Department of Radiology, SUNY Downstate Medical Center, Brooklyn, New York
| | - Jean Jeudy
- Department of Diagnostic Radiology and Nuclear Medicine, University of Maryland Medical Center, Baltimore, Maryland
| | - Alan Legasto
- Department of Radiology, Weill Cornell Medical College, New York, New York
| | - Stephen L Macknik
- Departments of Ophthalmology, Neurology, Physiology, and Pharmacology, SUNY Downstate Medical Center, Brooklyn, New York
| | - Susana Martinez-Conde
- Departments of Ophthalmology, Neurology, Physiology, and Pharmacology, SUNY Downstate Medical Center, Brooklyn, New York
| | | | - Deborah L Reede
- Department of Radiology, SUNY Downstate Medical Center, Brooklyn, New York
| |
Collapse
|
31
|
Littlefair S, Brennan P, Reed W, Mello-Thoms C. Does Expectation of Abnormality Affect the Search Pattern of Radiologists When Looking for Pulmonary Nodules? J Digit Imaging 2016; 30:55-62. [PMID: 27659798 DOI: 10.1007/s10278-016-9908-7] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022] Open
Abstract
This experiment investigated whether there might be an effect on the visual search strategy of radiologists during image interpretation of the same adult chest radiographs when given different clinical information. Each of 17 experienced radiologists was asked to interpret a set of 57 (10 abnormal) posteroanterior chest images to identify the presence of pulmonary lesions using differing clinical information (leading to unknown, low and high expectations of prevalence). Eye position metrics (search time, dwell time and time to first fixation) were compared for normal and abnormal images, as well as between conditions. For all images, there was a significantly longer search time at high prevalence expectation compared to low prevalence expectation (W = 75.19, P = <0.0001). Mann-Whitney analysis of the abnormal images demonstrated that the dwell time on correctly identified lesions was significantly shorter at low prevalence expectation compared to both unknown (U = 364.5, P = 0.02) and high prevalence expectation (U = 397.0, P = 0.0002). Visual search patterns of radiologists appear to be affected by changing a priori information where such information fosters an expectation of abnormality.
Collapse
Affiliation(s)
- Stephen Littlefair
- Medical Image Optimisation and Perception Group (MIOPeG), Discipline of Medical Radiation Sciences, Faculty of Health Sciences, University of Sydney, Room M213, 75, East Street, Sydney, NSW, 2141, Australia.
| | - Patrick Brennan
- Medical Image Optimisation and Perception Group (MIOPeG), Discipline of Medical Radiation Sciences, Faculty of Health Sciences, University of Sydney, Room M213, 75, East Street, Sydney, NSW, 2141, Australia
| | - Warren Reed
- Medical Image Optimisation and Perception Group (MIOPeG), Discipline of Medical Radiation Sciences, Faculty of Health Sciences, University of Sydney, Room M213, 75, East Street, Sydney, NSW, 2141, Australia
| | - Claudia Mello-Thoms
- Medical Image Optimisation and Perception Group (MIOPeG), Discipline of Medical Radiation Sciences, Faculty of Health Sciences, University of Sydney, Room M213, 75, East Street, Sydney, NSW, 2141, Australia
| |
Collapse
|
32
|
Feasibility of utilizing a commercial eye tracker to assess electronic health record use during patient simulation. Health Informatics J 2016; 22:744-57. [DOI: 10.1177/1460458215590250] [Citation(s) in RCA: 26] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Numerous reports describe unintended consequences of electronic health record implementation. Having previously described physicians’ failures to recognize patient safety issues within our electronic health record simulation environment, we now report on our use of eye and screen-tracking technology to understand factors associated with poor error recognition during an intensive care unit–based electronic health record simulation. We linked performance on the simulation to standard eye and screen-tracking readouts including number of fixations, saccades, mouse clicks and screens visited. In addition, we developed an overall Composite Eye Tracking score which measured when, where and how often each safety item was viewed. For 39 participants, the Composite Eye Tracking score correlated with performance on the simulation (p = 0.004). Overall, the improved performance was associated with a pattern of rapid scanning of data manifested by increased number of screens visited (p = 0.001), mouse clicks (p = 0.03) and saccades (p = 0.004). Eye tracking can be successfully integrated into electronic health record–based simulation and provides a surrogate measure of cognitive decision making and electronic health record usability.
Collapse
|
33
|
Fong A, Hoffman DJ, Zachary Hettinger A, Fairbanks RJ, Bisantz AM. Identifying visual search patterns in eye gaze data; gaining insights into physician visual workflow. J Am Med Inform Assoc 2016; 23:1180-1184. [DOI: 10.1093/jamia/ocv196] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2015] [Revised: 09/28/2015] [Accepted: 11/10/2015] [Indexed: 11/12/2022] Open
Abstract
Abstract
Importance and Objectives As health information technologies become more prevalent in physician workflow, it is increasingly important to understand how physicians are using and interacting with these systems. This includes understanding how physicians search for information presented through health information technology systems. Eye tracking technologies provide a useful technique to understand how physicians visually search for information. However, analyzing eye tracking data can be challenging and is often done by measuring summative metrics, such as total time looking at a specific area and first-order transitions.
Methods In this paper, we propose an algorithmic approach to identify different visual search patterns. We demonstrate this approach by identifying common visual search patterns from physicians using a simulated prototype emergency department patient tracking system.
Results and Conclusions We evaluate and compare the visual search pattern results to first-order transition results. We discuss the benefits and limitations of this approach and insights from this initial evaluation.
Collapse
Affiliation(s)
- Allan Fong
- MedStar Institute for Innovation - National Center for Human Factors in Healthcare, 3007 Tilden St. NW, Suite 7M, Washington, DC, 20008, USA
| | - Daniel J Hoffman
- MedStar Institute for Innovation - National Center for Human Factors in Healthcare, 3007 Tilden St. NW, Suite 7M, Washington, DC, 20008, USA
| | - A Zachary Hettinger
- MedStar Institute for Innovation - National Center for Human Factors in Healthcare, 3007 Tilden St. NW, Suite 7M, Washington, DC, 20008, USA
- Georgetown University School of Medicine, 3800 Reservoir Rd NW, Washington, DC 20007
| | - Rollin J Fairbanks
- MedStar Institute for Innovation - National Center for Human Factors in Healthcare, 3007 Tilden St. NW, Suite 7M, Washington, DC, 20008, USA
- Georgetown University School of Medicine, 3800 Reservoir Rd NW, Washington, DC 20007
- The State University of New York, University at Buffalo, Amherst, NY 14216, USA
| | - Ann M Bisantz
- The State University of New York, University at Buffalo, Amherst, NY 14216, USA
| |
Collapse
|
34
|
Screen-Based Simulation, Virtual Reality, and Haptic Simulators. COMPREHENSIVE HEALTHCARE SIMULATION: PEDIATRICS 2016. [DOI: 10.1007/978-3-319-24187-6_9] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/13/2023]
|
35
|
Darcy S, Rainford L, Kelly B, Toomey R. Decision Making and Variation in Radiation Exposure Factor Selection by Radiologic Technologists. J Med Imaging Radiat Sci 2015; 46:372-379. [PMID: 31052117 DOI: 10.1016/j.jmir.2015.09.003] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2015] [Revised: 09/11/2015] [Accepted: 09/11/2015] [Indexed: 01/27/2023]
Abstract
The goal of radiographic imaging is to produce a diagnostically useful image while minimizing patient radiation dose. This study aimed to review variations in exposure factor selection by radiologic technologists for virtual patients with varying body mass index characteristics. Eleven technologists were asked to assign exposure parameters (kVp, mAs, source-to-image receptor distance, and grid use) to 10 computer-generated patient images for each of four radiographic examinations (anteroposterior [AP] shoulder; AP lumbar spine; lateral lumbar spine; AP portable chest). The virtual patients represented five body mass index categories-underweight, healthy weight, overweight, obese, and superobese. As participants assigned exposures, their visual patterns were recorded by a Tobii TX300 eye-tracker. Significant (P < .05) correlation was found between radiographer age/experience and assignment of mAs for AP shoulder and lumbar examinations. Greater age/experience correlated with higher mAs for the AP shoulder examination, but with lower values for lumbar examinations. Strong correlations also existed between times to first fixations on relevant anatomic areas, and kVp/mAs values existed for the AP portable chest examination. Exposure selection differences related to age/experience highlight inconsistencies in the practice of exposure parameter setting. The reason for these inconsistencies requires further investigation, and how to address deficiencies in practice requires consideration to optimize safe patient care. Because of the small sample size used, further research into the relationship between visual factors and individual examinations is suggested, after the findings regarding the AP portable chest examination.
Collapse
Affiliation(s)
- Sarah Darcy
- School of Medicine and Medical Science, University College Dublin, Dublin 4, Ireland.
| | - Louise Rainford
- Department of Diagnostic Imaging, School of Medicine and Medical Science, University College Dublin, Dublin 4, Ireland
| | - Brendan Kelly
- School of Medicine and Medical Science, University College Dublin, Dublin 4, Ireland
| | - Rachel Toomey
- Department of Diagnostic Imaging, School of Medicine and Medical Science, University College Dublin, Dublin 4, Ireland
| |
Collapse
|
36
|
Doberne JW, He Z, Mohan V, Gold JA, Marquard J, Chiang MF. Using High-Fidelity Simulation and Eye Tracking to Characterize EHR Workflow Patterns among Hospital Physicians. AMIA ... ANNUAL SYMPOSIUM PROCEEDINGS. AMIA SYMPOSIUM 2015; 2015:1881-1889. [PMID: 26958287 PMCID: PMC4765617] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
Modern EHR systems are complex, and end-user behavior and training are highly variable. The need for clinicians to access key clinical data is a critical patient safety issue. This study used a mixed methods approach employing a high-fidelity EHR simulation environment, eye and screen tracking, surveys, and semi-structured interviews to characterize typical EHR usage by hospital physicians (hospitalists) as they encounter a new patient. The main findings were: 1) There were strong similarities across the groups in the information types the physicians looked at most frequently, 2) While there was no overall difference in case duration between the groups, we observed two distinct workflow types between the groups with respect to gathering information in the EHR and creating a note, and 3) A majority of the case time was devoted to note composition in both groups. This has implications for EHR interface design and raises further questions about what individual user workflows exist in the EHR.
Collapse
Affiliation(s)
- Julie W Doberne
- Department of Medical Informatics & Clinical Epidemiology, Oregon Health & Science University, Portland, OR
| | - Ze He
- University of Massachusetts Amherst, Amherst, MA
| | - Vishnu Mohan
- Department of Medical Informatics & Clinical Epidemiology, Oregon Health & Science University, Portland, OR
| | - Jeffrey A Gold
- Department of Medical Informatics & Clinical Epidemiology, Oregon Health & Science University, Portland, OR; Department of Pulmonology & Critical Care, Oregon Health & Science University, Portland, OR
| | | | - Michael F Chiang
- Department of Medical Informatics & Clinical Epidemiology, Oregon Health & Science University, Portland, OR; Department of Ophthalmology, Oregon Health & Science University, Portland, OR
| |
Collapse
|
37
|
Han D, Wang S, Jiang C, Jiang X, Kim HE, Sun J, Ohno-Machado L. Trends in biomedical informatics: automated topic analysis of JAMIA articles. J Am Med Inform Assoc 2015; 22:1153-63. [PMID: 26555018 PMCID: PMC5009912 DOI: 10.1093/jamia/ocv157] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2015] [Revised: 09/08/2015] [Accepted: 09/14/2015] [Indexed: 01/26/2023] Open
Abstract
Biomedical Informatics is a growing interdisciplinary field in which research topics and citation trends have been evolving rapidly in recent years. To analyze these data in a fast, reproducible manner, automation of certain processes is needed. JAMIA is a "generalist" journal for biomedical informatics. Its articles reflect the wide range of topics in informatics. In this study, we retrieved Medical Subject Headings (MeSH) terms and citations of JAMIA articles published between 2009 and 2014. We use tensors (i.e., multidimensional arrays) to represent the interaction among topics, time and citations, and applied tensor decomposition to automate the analysis. The trends represented by tensors were then carefully interpreted and the results were compared with previous findings based on manual topic analysis. A list of most cited JAMIA articles, their topics, and publication trends over recent years is presented. The analyses confirmed previous studies and showed that, from 2012 to 2014, the number of articles related to MeSH terms Methods, Organization & Administration, and Algorithms increased significantly both in number of publications and citations. Citation trends varied widely by topic, with Natural Language Processing having a large number of citations in particular years, and Medical Record Systems, Computerized remaining a very popular topic in all years.
Collapse
Affiliation(s)
- Dong Han
- Health System Department of Biomedical Informatics, University of California San Diego, La Jolla, CA, 92093, USA School of Electrical and Computer Engineering, University of Oklahoma, Tulsa, OK, 74135, USA
| | - Shuang Wang
- Health System Department of Biomedical Informatics, University of California San Diego, La Jolla, CA, 92093, USA
| | - Chao Jiang
- Health System Department of Biomedical Informatics, University of California San Diego, La Jolla, CA, 92093, USA School of Electrical and Computer Engineering, University of Oklahoma, Tulsa, OK, 74135, USA
| | - Xiaoqian Jiang
- Health System Department of Biomedical Informatics, University of California San Diego, La Jolla, CA, 92093, USA
| | - Hyeon-Eui Kim
- Health System Department of Biomedical Informatics, University of California San Diego, La Jolla, CA, 92093, USA
| | - Jimeng Sun
- School of Computational Science and Engineering, Georgia Institute of Technology, Atlanta, GA, S30313, USA
| | - Lucila Ohno-Machado
- Health System Department of Biomedical Informatics, University of California San Diego, La Jolla, CA, 92093, USA
| |
Collapse
|
38
|
Zamacona JR, Niehaus R, Rasin A, Furst JD, Raicu DS. Assessing diagnostic complexity: An image feature-based strategy to reduce annotation costs. Comput Biol Med 2015; 62:294-305. [DOI: 10.1016/j.compbiomed.2015.01.013] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2014] [Revised: 01/05/2015] [Accepted: 01/14/2015] [Indexed: 11/26/2022]
|
39
|
Zhang J, Lo JY, Kuzmiak CM, Ghate SV, Yoon SC, Mazurowski MA. Using computer-extracted image features for modeling of error-making patterns in detection of mammographic masses among radiology residents. Med Phys 2015; 41:091907. [PMID: 25186394 DOI: 10.1118/1.4892173] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/13/2022] Open
Abstract
PURPOSE Mammography is the most widely accepted and utilized screening modality for early breast cancer detection. Providing high quality mammography education to radiology trainees is essential, since excellent interpretation skills are needed to ensure the highest benefit of screening mammography for patients. The authors have previously proposed a computer-aided education system based on trainee models. Those models relate human-assessed image characteristics to trainee error. In this study, the authors propose to build trainee models that utilize features automatically extracted from images using computer vision algorithms to predict likelihood of missing each mass by the trainee. This computer vision-based approach to trainee modeling will allow for automatically searching large databases of mammograms in order to identify challenging cases for each trainee. METHODS The authors' algorithm for predicting the likelihood of missing a mass consists of three steps. First, a mammogram is segmented into air, pectoral muscle, fatty tissue, dense tissue, and mass using automated segmentation algorithms. Second, 43 features are extracted using computer vision algorithms for each abnormality identified by experts. Third, error-making models (classifiers) are applied to predict the likelihood of trainees missing the abnormality based on the extracted features. The models are developed individually for each trainee using his/her previous reading data. The authors evaluated the predictive performance of the proposed algorithm using data from a reader study in which 10 subjects (7 residents and 3 novices) and 3 experts read 100 mammographic cases. Receiver operating characteristic (ROC) methodology was applied for the evaluation. RESULTS The average area under the ROC curve (AUC) of the error-making models for the task of predicting which masses will be detected and which will be missed was 0.607 (95% CI,0.564-0.650). This value was statistically significantly different from 0.5 (p<0.0001). For the 7 residents only, the AUC performance of the models was 0.590 (95% CI,0.537-0.642) and was also significantly higher than 0.5 (p=0.0009). Therefore, generally the authors' models were able to predict which masses were detected and which were missed better than chance. CONCLUSIONS The authors proposed an algorithm that was able to predict which masses will be detected and which will be missed by each individual trainee. This confirms existence of error-making patterns in the detection of masses among radiology trainees. Furthermore, the proposed methodology will allow for the optimized selection of difficult cases for the trainees in an automatic and efficient manner.
Collapse
Affiliation(s)
- Jing Zhang
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina 27705
| | - Joseph Y Lo
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina 27705; Duke Cancer Institute, Durham, North Carolina 27710; Departments of Biomedical Engineering and Electrical & Computer Engineering, Duke University, Durham, North Carolina 27705; and Medical Physics Graduate Program, Duke University, Durham, North Carolina 27705
| | - Cherie M Kuzmiak
- Department of Radiology, University of North Carolina at Chapel Hill School of Medicine, Chapel Hill, North Carolina 27599
| | - Sujata V Ghate
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina 27705
| | - Sora C Yoon
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina 27705
| | - Maciej A Mazurowski
- Department of Radiology, Duke University School of Medicine, Durham, North Carolina 27705; Duke Cancer Institute, Durham, North Carolina 27710; and Medical Physics Graduate Program, Duke University, Durham, North Carolina 27705
| |
Collapse
|
40
|
Zhang J, Silber JI, Mazurowski MA. Modeling false positive error making patterns in radiology trainees for improved mammography education. J Biomed Inform 2015; 54:50-7. [PMID: 25640462 DOI: 10.1016/j.jbi.2015.01.007] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2014] [Revised: 01/13/2015] [Accepted: 01/19/2015] [Indexed: 10/24/2022]
Abstract
INTRODUCTION While mammography notably contributes to earlier detection of breast cancer, it has its limitations, including a large number of false positive exams. Improved radiology education could potentially contribute to alleviating this issue. Toward this goal, in this paper we propose an algorithm for modeling of false positive error making among radiology trainees. Identifying troublesome locations for the trainees could focus their training and in turn improve their performance. METHODS The algorithm proposed in this paper predicts locations that are likely to result in a false positive error for each trainee based on the previous annotations made by the trainee. The algorithm consists of three steps. First, the suspicious false positive locations are identified in mammograms by Difference of Gaussian filter and suspicious regions are segmented by computer vision-based segmentation algorithms. Second, 133 features are extracted for each suspicious region to describe its distinctive characteristics. Third, a random forest classifier is applied to predict the likelihood of the trainee making a false positive error using the extracted features. The random forest classifier is trained using previous annotations made by the trainee. We evaluated the algorithm using data from a reader study in which 3 experts and 10 trainees interpreted 100 mammographic cases. RESULTS The algorithm was able to identify locations where the trainee will commit a false positive error with accuracy higher than an algorithm that selects such locations randomly. Specifically, our algorithm found false positive locations with 40% accuracy when only 1 location was selected for all cases for each trainee and 12% accuracy when 10 locations were selected. The accuracies for randomly identified locations were both 0% for these two scenarios. CONCLUSIONS In this first study on the topic, we were able to build computer models that were able to find locations for which a trainee will make a false positive error in images that were not previously seen by the trainee. Presenting the trainees with such locations rather than randomly selected ones may improve their educational outcomes.
Collapse
Affiliation(s)
- Jing Zhang
- Department of Radiology, Duke University School of Medicine, Durham, NC, United States; Computer Science Department, Lamar University, Beaumont, TX, United States.
| | - James I Silber
- Department of Biomedical Engineering, Duke University Pratt School of Engineering, Durham, NC, United States
| | - Maciej A Mazurowski
- Department of Radiology, Duke University School of Medicine, Durham, NC, United States; Duke Cancer Institute, United States; Duke Medical Physics Program, United States
| |
Collapse
|
41
|
Rubin GD, Roos JE, Tall M, Harrawood B, Bag S, Ly DL, Seaman DM, Hurwitz LM, Napel S, Roy Choudhury K. Characterizing search, recognition, and decision in the detection of lung nodules on CT scans: elucidation with eye tracking. Radiology 2014; 274:276-86. [PMID: 25325324 DOI: 10.1148/radiol.14132918] [Citation(s) in RCA: 65] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
Abstract
PURPOSE To determine the effectiveness of radiologists' search, recognition, and acceptance of lung nodules on computed tomographic (CT) images by using eye tracking. MATERIALS AND METHODS This study was performed with a protocol approved by the institutional review board. All study subjects provided informed consent, and all private health information was protected in accordance with HIPAA. A remote eye tracker was used to record time-varying gaze paths while 13 radiologists interpreted 40 lung CT images with an average of 3.9 synthetic nodules (5-mm diameter) embedded randomly in the lung parenchyma. The radiologists' gaze volumes ( GV gaze volume s) were defined as the portion of the lung parenchyma within 50 pixels (approximately 3 cm) of all gaze points. The fraction of the total lung volume encompassed within the GV gaze volume s, the fraction of lung nodules encompassed within each GV gaze volume (search effectiveness), the fraction of lung nodules within the GV gaze volume detected by the reader (recognition-acceptance effectiveness), and overall sensitivity of lung nodule detection were measured. RESULTS Detected nodules were within 50 pixels of the nearest gaze point for 990 of 992 correct detections. On average, radiologists searched 26.7% of the lung parenchyma in 3 minutes and 16 seconds and encompassed between 86 and 143 of 157 nodules within their GV gaze volume s. Once encompassed within their GV gaze volume , the average sensitivity of nodule recognition and acceptance ranged from 47 of 100 nodules to 103 of 124 nodules (sensitivity, 0.47-0.82). Overall sensitivity ranged from 47 to 114 of 157 nodules (sensitivity, 0.30-0.73) and showed moderate correlation (r = 0.62, P = .02) with the fraction of lung volume searched. CONCLUSION Relationships between reader search, recognition and acceptance, and overall lung nodule detection rate can be studied with eye tracking. Radiologists appear to actively search less than half of the lung parenchyma, with substantial interreader variation in volume searched, fraction of nodules included within the search volume, sensitivity for nodules within the search volume, and overall detection rate.
Collapse
Affiliation(s)
- Geoffrey D Rubin
- From the Duke Clinical Research Institute, Box 17969, 2400 Pratt St, Durham, NC 27715 (G.D.R., K.R.C.); Department of Radiology, Duke University School of Medicine, Durham, NC (G.D.R., J.E.R., M.T., B.H., S.B., D.M.S., L.M.H.); Department of Medical Imaging, University of Toronto, Toronto, ON, Canada (D.L.L.); and Department of Radiology, Stanford University School of Medicine, Stanford, Calif (S.N.)
| | | | | | | | | | | | | | | | | | | |
Collapse
|
42
|
Depeursinge A, Kurtz C, Beaulieu CF, Napel S, Rubin DL. Predicting visual semantic descriptive terms from radiological image data: preliminary results with liver lesions in CT. IEEE TRANSACTIONS ON MEDICAL IMAGING 2014; 33:1669-76. [PMID: 24808406 PMCID: PMC4129229 DOI: 10.1109/tmi.2014.2321347] [Citation(s) in RCA: 25] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
We describe a framework to model visual semantics of liver lesions in CT images in order to predict the visual semantic terms (VST) reported by radiologists in describing these lesions. Computational models of VST are learned from image data using linear combinations of high-order steerable Riesz wavelets and support vector machines (SVM). In a first step, these models are used to predict the presence of each semantic term that describes liver lesions. In a second step, the distances between all VST models are calculated to establish a nonhierarchical computationally-derived ontology of VST containing inter-term synonymy and complementarity. A preliminary evaluation of the proposed framework was carried out using 74 liver lesions annotated with a set of 18 VSTs from the RadLex ontology. A leave-one-patient-out cross-validation resulted in an average area under the ROC curve of 0.853 for predicting the presence of each VST. The proposed framework is expected to foster human-computer synergies for the interpretation of radiological images while using rotation-covariant computational models of VSTs to 1) quantify their local likelihood and 2) explicitly link them with pixel-based image content in the context of a given imaging domain.
Collapse
Affiliation(s)
- Adrien Depeursinge
- Department of Radiology of the School of Medicine, Stanford University, CA, USA
| | - Camille Kurtz
- Department of Radiology of the School of Medicine, Stanford University, CA, USA
- C. Kurtz is also with the LIPADE (EA2517), University Paris Descartes, France
| | | | - Sandy Napel
- Department of Radiology of the School of Medicine, Stanford University, CA, USA
| | - Daniel L. Rubin
- Department of Radiology of the School of Medicine, Stanford University, CA, USA
| |
Collapse
|
43
|
Grimm LJ, Kuzmiak CM, Ghate SV, Yoon SC, Mazurowski MA. Radiology resident mammography training: interpretation difficulty and error-making patterns. Acad Radiol 2014; 21:888-92. [PMID: 24928157 DOI: 10.1016/j.acra.2014.01.025] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2013] [Revised: 01/20/2014] [Accepted: 01/24/2014] [Indexed: 11/25/2022]
Abstract
RATIONALE AND OBJECTIVES The purpose of this study was to better understand the concept of mammography difficulty and how it affects radiology resident performance. MATERIALS AND METHODS Seven radiology residents and three expert breast imagers reviewed 100 mammograms, consisting of bilateral medial lateral oblique and craniocaudal views, using a research workstation. The cases consisted of normal, benign, and malignant findings. Participants identified abnormalities and scored the difficulty and malignant potential for each case. Resident performance (sensitivity, specificity, and area under the receiver operating characteristic curve [AUC]) was calculated for self- and expert-assessed high and low difficulties. RESULTS For cases classified by self-assessed difficulty, the resident AUCs were 0.667 for high difficulty and 0.771 for low difficulty cases (P = .010). Resident sensitivities were 0.707 for high and 0.614 for low difficulty cases (P = .113). Resident specificities were 0.583 for high and 0.905 for low difficulty cases (P < .001). For cases classified by expert-assessed difficulty, the resident AUCs were 0.583 for high and 0.783 for low difficulty cases (P = .001). Resident sensitivities were 0.558 for high and 0.796 for low difficulty cases (P < .001). Resident specificities were 0.714 for high and 0.740 for low difficulty cases (P = .807). CONCLUSIONS Increased self- and expert-assessed difficulty is associated with a decrease in resident performance in mammography. However, while this lower performance is due to a decrease in specificity for self-assessed difficulty, it is due to a decrease in sensitivity for expert-assessed difficulty. These trends suggest that educators should provide a mix of self- and expert-assessed difficult cases in educational materials to maximize the effect of training on resident performance and confidence.
Collapse
|
44
|
Grimm LJ, Ghate SV, Yoon SC, Kuzmiak CM, Kim C, Mazurowski MA. Predicting error in detecting mammographic masses among radiology trainees using statistical models based on BI-RADS features. Med Phys 2014; 41:031909. [DOI: 10.1118/1.4866379] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/24/2022] Open
|
45
|
Hsu W, Markey MK, Wang MD. Biomedical imaging informatics in the era of precision medicine: progress, challenges, and opportunities. J Am Med Inform Assoc 2013; 20:1010-3. [PMID: 24114330 DOI: 10.1136/amiajnl-2013-002315] [Citation(s) in RCA: 30] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022] Open
Affiliation(s)
- William Hsu
- Medical Imaging Informatics (MII) Group, Department of Radiological Sciences, UCLA David Geffen School of Medicine, Los Angeles, California, USA
| | | | | |
Collapse
|