1
|
Awasthi A, Le N, Deng Z, Agrawal R, Wu CC, Van Nguyen H. Bridging human and machine intelligence: Reverse-engineering radiologist intentions for clinical trust and adoption. Comput Struct Biotechnol J 2024; 24:711-723. [PMID: 39660015 PMCID: PMC11629193 DOI: 10.1016/j.csbj.2024.11.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2024] [Revised: 10/19/2024] [Accepted: 11/04/2024] [Indexed: 12/12/2024] Open
Abstract
In the rapidly evolving landscape of medical imaging, the integration of artificial intelligence (AI) with clinical expertise offers unprecedented opportunities to enhance diagnostic precision and accuracy. Yet, the "black box" nature of AI models often limits their integration into clinical practice, where transparency and interpretability are important. This paper presents a novel system leveraging the Large Multimodal Model (LMM) to bridge the gap between AI predictions and the cognitive processes of radiologists. This system consists of two core modules, Temporally Grounded Intention Detection (TGID) and Region Extraction (RE). The TGID module predicts the radiologist's intentions by analyzing eye gaze fixation heatmap videos and corresponding radiology reports. Additionally, the RE module extracts regions of interest that align with these intentions, mirroring the radiologist's diagnostic focus. This approach introduces a new task, radiologist intention detection, and is the first application of Dense Video Captioning (DVC) in the medical domain. By making AI systems more interpretable and aligned with radiologist's cognitive processes, this proposed system aims to enhance trust, improve diagnostic accuracy, and support medical education. Additionally, it holds the potential for automated error correction, guiding junior radiologists, and fostering more effective training and feedback mechanisms. This work sets a precedent for future research in AI-driven healthcare, offering a pathway towards transparent, trustworthy, and human-centered AI systems. We evaluated this model using NLG(Natural Language Generation), time-related, and vision-based metrics, demonstrating superior performance in generating temporally grounded intentions on REFLACX and EGD-CXR datasets. This model also demonstrated strong predictive accuracy in overlap scores for medical abnormalities and effective region extraction with high IoU(Intersection over Union), especially in complex cases like cardiomegaly and edema. These results highlight the system's potential to enhance diagnostic accuracy and support continuous learning in radiology.
Collapse
Affiliation(s)
- Akash Awasthi
- Department of Electrical and Computer Engineering, University of Houston, United States
| | - Ngan Le
- Department of Computer Science & Computer Engineering, University of Arkansas, United States
| | - Zhigang Deng
- Department of Computer Science, University of Houston, Houston, TX, United States
| | - Rishi Agrawal
- Department of Thoracic Imaging, Division of Diagnostic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Carol C. Wu
- Department of Thoracic Imaging, Division of Diagnostic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Hien Van Nguyen
- Department of Electrical and Computer Engineering, University of Houston, United States
| |
Collapse
|
2
|
Kulkarni S, Singh S, Nagi A, Gulhane A. Dissonance Between Law Courts and the Science of Visual Perception in Medical Imaging. Acad Radiol 2024:S1076-6332(24)00574-9. [PMID: 39389813 DOI: 10.1016/j.acra.2024.08.011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/19/2024] [Revised: 07/31/2024] [Accepted: 08/01/2024] [Indexed: 10/12/2024]
Abstract
Lawsuits for spending too little time interpreting each radiological image are a vexatious charge to level against a radiologist in medical malpractice court. In this article, we recount two medicolegal cases where the defendant radiologists were accused of missing a life-threatening diagnosis due to not spending enough time reviewing each image. We consider the literature in vision sciences, visual perception in radiology and interpretive biases to demonstrate that using reading speed as evidence of negligence in a malpractice court represents in incorrect understanding of how radiologists perceive images, including three-dimensional volumetric studies.
Collapse
Affiliation(s)
- Sagar Kulkarni
- Department of Radiology, University of Washington Medical Center, Seattle, Washington, USA (S.K., S.S., A.G.).
| | - Sarabjeet Singh
- Department of Radiology, University of Washington Medical Center, Seattle, Washington, USA (S.K., S.S., A.G.)
| | - Ajeet Nagi
- Department of Biomedical Engineering, The University of Texas at Austin, Austin, Texas, USA (A.N.)
| | - Avanti Gulhane
- Department of Radiology, University of Washington Medical Center, Seattle, Washington, USA (S.K., S.S., A.G.)
| |
Collapse
|
3
|
Byrne CA, Voute LC, Marshall JF. Interobserver agreement during clinical magnetic resonance imaging of the equine foot. Equine Vet J 2024. [PMID: 38946165 DOI: 10.1111/evj.14126] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2023] [Accepted: 06/02/2024] [Indexed: 07/02/2024]
Abstract
BACKGROUND Agreement between experienced observers for assessment of pathology and assessment confidence are poorly documented for magnetic resonance imaging (MRI) of the equine foot. OBJECTIVES To report interobserver agreement for pathology assessment and observer confidence for key anatomical structures of the equine foot during MRI. STUDY DESIGN Exploratory clinical study. METHODS Ten experienced observers (diploma or associate level) assessed 15 equine foot MRI studies acquired from clinical databases of 3 MRI systems. Observers graded pathology in seven key anatomical structures (Grade 1: no pathology, Grade 2: mild pathology, Grade 3: moderate pathology, Grade 4: severe pathology) and provided a grade for their confidence for each pathology assessment (Grade 1: high confidence, Grade 2: moderate confidence, Grade 3: limited confidence, Grade 4: no confidence). Interobserver agreement for the presence/absence of pathology and agreement for individual grades of pathology were assessed with Fleiss' kappa (k). Overall interobserver agreement for pathology was determined using Fleiss' kappa and Kendall's coefficient of concordance (KCC). The distribution of grading was also visualised with bubble charts. RESULTS Interobserver agreement for the presence/absence of pathology of individual anatomical structures was poor-to-fair, except for the navicular bone which had moderate agreement (k = 0.52). Relative agreement for pathology grading (accounting for the ranking of grades) ranged from KCC = 0.19 for the distal interphalangeal joint to KCC = 0.70 for the navicular bone. Agreement was generally greatest at the extremes of pathology. Observer confidence in pathology assessment was generally moderate to high. MAIN LIMITATIONS Distribution of pathology varied between anatomical structures due to random selection of clinical MRI studies. Observers had most experience with low-field MRI. CONCLUSIONS Even with experienced observers, there can be notable variation in the perceived severity of foot pathology on MRI for individual cases, which could be important in a clinical context.
Collapse
Affiliation(s)
- Christian A Byrne
- School of Veterinary Medicine, College of Medical, Veterinary and Life Sciences, University of Glasgow, Glasgow, UK
| | - Lance C Voute
- School of Veterinary Medicine, College of Medical, Veterinary and Life Sciences, University of Glasgow, Glasgow, UK
| | - John F Marshall
- School of Veterinary Medicine, College of Medical, Veterinary and Life Sciences, University of Glasgow, Glasgow, UK
| |
Collapse
|
4
|
Alexander RG, Venkatakrishnan A, Chanovas J, Ferguson S, Macknik SL, Martinez-Conde S. Why did Rubens add a parrot to Titian's The Fall of Man? A pictorial manipulation of joint attention. J Vis 2024; 24:1. [PMID: 38558160 PMCID: PMC10996941 DOI: 10.1167/jov.24.4.1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2023] [Accepted: 01/19/2024] [Indexed: 04/04/2024] Open
Abstract
Almost 400 years ago, Rubens copied Titian's The Fall of Man, albeit with important changes. Rubens altered Titian's original composition in numerous ways, including by changing the gaze directions of the depicted characters and adding a striking red parrot to the painting. Here, we quantify the impact of Rubens's choices on the viewer's gaze behavior. We displayed digital copies of Rubens's and Titian's artworks-as well as a version of Rubens's painting with the parrot digitally removed-on a computer screen while recording the eye movements produced by observers during free visual exploration of each image. To assess the effects of Rubens's changes to Titian's composition, we directly compared multiple gaze parameters across the different images. We found that participants gazed at Eve's face more frequently in Rubens's painting than in Titian's. In addition, gaze positions were more tightly focused for the former than for the latter, consistent with different allocations of viewer interest. We also investigated how gaze fixation on Eve's face affected the perceptual visibility of the parrot in Rubens's composition and how the parrot's presence versus its absence impacted gaze dynamics. Taken together, our results demonstrate that Rubens's critical deviations from Titian's painting have powerful effects on viewers' oculomotor behavior.
Collapse
Affiliation(s)
- Robert G Alexander
- Department of Psychology & Counseling, New York Institute of Technology, New York, NY, USA
| | - Ashwin Venkatakrishnan
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Jordi Chanovas
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
- Graduate Program in Neural and Behavioral Science, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Sophie Ferguson
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Stephen L Macknik
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Susana Martinez-Conde
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| |
Collapse
|
5
|
He Q, Zhu X, Fang F. Enhancing visual perceptual learning using transcranial electrical stimulation: Transcranial alternating current stimulation outperforms both transcranial direct current and random noise stimulation. J Vis 2023; 23:2. [PMID: 38054934 PMCID: PMC10702794 DOI: 10.1167/jov.23.14.2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2023] [Accepted: 10/23/2023] [Indexed: 12/07/2023] Open
Abstract
Diverse strategies can be employed to enhance visual skills, including visual perceptual learning (VPL) and transcranial electrical stimulation (tES). Combining VPL and tES is a popular method that holds promise for producing significant improvements in visual acuity within a short time frame. However, there is still a lack of comprehensive evaluation regarding the effects of combining different types of tES and VPL on enhancing visual function, especially with a larger sample size. In the present study, we recruited four groups of subjects (26 subjects each) to learn an orientation discrimination task with five daily training sessions. During training, the occipital region of each subject was stimulated by one type of tES-anodal transcranial direct current stimulation (tDCS), alternating current stimulation (tACS) at 10 Hz, high-frequency random noise stimulation (tRNS), and sham tACS-while the subject performed the training task. We found that, compared with the sham stimulation, both the high-frequency tRNS and the 10-Hz tACS facilitated VPL efficiently in terms of learning rate and performance improvement, but there was little modulatory effect in the anodal tDCS condition. Remarkably, the 10-Hz tACS condition exhibited superior modulatory effects compared with the tRNS condition, demonstrating the strongest modulation among the most commonly used tES types for further enhancing vision when combined with VPL. Our results suggest that alpha oscillations play a vital role in VPL. Our study provides a practical guide for vision rehabilitation.
Collapse
Affiliation(s)
- Qing He
- School of Psychological and Cognitive Sciences and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing, China
- Key Laboratory of Machine Perception, Ministry of Education, Peking University, Beijing, China
- IDG/McGovern Institute for Brain Research, Peking University, Beijing, China
- Peking-Tsinghua Center for Life Sciences, Peking University, Beijing, China
| | - Xinyi Zhu
- School of Psychological and Cognitive Sciences and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing, China
- Key Laboratory of Machine Perception, Ministry of Education, Peking University, Beijing, China
- IDG/McGovern Institute for Brain Research, Peking University, Beijing, China
- Peking-Tsinghua Center for Life Sciences, Peking University, Beijing, China
| | - Fang Fang
- School of Psychological and Cognitive Sciences and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing, China
- Key Laboratory of Machine Perception, Ministry of Education, Peking University, Beijing, China
- IDG/McGovern Institute for Brain Research, Peking University, Beijing, China
- Peking-Tsinghua Center for Life Sciences, Peking University, Beijing, China
| |
Collapse
|
6
|
Borgbjerg J, Thompson JD, Salte IM, Frøkjær JB. Towards AI-augmented radiology education: a web-based application for perception training in chest X-ray nodule detection. Br J Radiol 2023; 96:20230299. [PMID: 37750851 PMCID: PMC10646630 DOI: 10.1259/bjr.20230299] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2023] [Revised: 07/09/2023] [Accepted: 08/15/2023] [Indexed: 09/27/2023] Open
Abstract
OBJECTIVES Artificial intelligence (AI)-based applications for augmenting radiological education are underexplored. Prior studies have demonstrated the effectiveness of simulation in radiological perception training. This study aimed to develop and make available a pure web-based application called Perception Trainer for perception training in lung nodule detection in chest X-rays. METHODS Based on open-access data, we trained a deep-learning model for lung segmentation in chest X-rays. Subsequently, an algorithm for artificial lung nodule generation was implemented and combined with the segmentation model to allow on-the-fly procedural insertion of lung nodules in chest X-rays. This functionality was integrated into an existing zero-footprint web-based DICOM viewer, and a dynamic HTML page was created to specify case generation parameters. RESULTS The result is an easily accessible platform-agnostic web application available at: https://castlemountain.dk/mulrecon/perceptionTrainer.html.The application allows the user to specify the characteristics of lung nodules to be inserted into chest X-rays, and it produces automated feedback regarding nodule detection performance. Generated cases can be shared through a uniform resource locator. CONCLUSION We anticipate that the description and availability of our developed solution with open-sourced codes may help facilitate radiological education and stimulate the development of similar AI-augmented educational tools. ADVANCES IN KNOWLEDGE A web-based application applying AI-based techniques for radiological perception training was developed. The application demonstrates a novel approach for on-the-fly generation of cases in chest X-ray lung nodule detection employing deep-learning-based segmentation and lung nodule simulation.
Collapse
Affiliation(s)
- Jens Borgbjerg
- Department of Radiology, Akershus University Hospital, Oslo, Norway
| | - John D Thompson
- Department of Radiology, University Hospitals of Morecambe Bay NHS Foundation Trust, Morecambe, United Kingdom
| | | | | |
Collapse
|
7
|
Holmqvist K, Örbom SL, Hooge ITC, Niehorster DC, Alexander RG, Andersson R, Benjamins JS, Blignaut P, Brouwer AM, Chuang LL, Dalrymple KA, Drieghe D, Dunn MJ, Ettinger U, Fiedler S, Foulsham T, van der Geest JN, Hansen DW, Hutton SB, Kasneci E, Kingstone A, Knox PC, Kok EM, Lee H, Lee JY, Leppänen JM, Macknik S, Majaranta P, Martinez-Conde S, Nuthmann A, Nyström M, Orquin JL, Otero-Millan J, Park SY, Popelka S, Proudlock F, Renkewitz F, Roorda A, Schulte-Mecklenbeck M, Sharif B, Shic F, Shovman M, Thomas MG, Venrooij W, Zemblys R, Hessels RS. Eye tracking: empirical foundations for a minimal reporting guideline. Behav Res Methods 2023; 55:364-416. [PMID: 35384605 PMCID: PMC9535040 DOI: 10.3758/s13428-021-01762-8] [Citation(s) in RCA: 55] [Impact Index Per Article: 27.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/29/2021] [Indexed: 11/08/2022]
Abstract
In this paper, we present a review of how the various aspects of any study using an eye tracker (such as the instrument, methodology, environment, participant, etc.) affect the quality of the recorded eye-tracking data and the obtained eye-movement and gaze measures. We take this review to represent the empirical foundation for reporting guidelines of any study involving an eye tracker. We compare this empirical foundation to five existing reporting guidelines and to a database of 207 published eye-tracking studies. We find that reporting guidelines vary substantially and do not match with actual reporting practices. We end by deriving a minimal, flexible reporting guideline based on empirical research (Section "An empirically based minimal reporting guideline").
Collapse
Affiliation(s)
- Kenneth Holmqvist
- Department of Psychology, Nicolaus Copernicus University, Torun, Poland.
- Department of Computer Science and Informatics, University of the Free State, Bloemfontein, South Africa.
- Department of Psychology, Regensburg University, Regensburg, Germany.
| | - Saga Lee Örbom
- Department of Psychology, Regensburg University, Regensburg, Germany
| | - Ignace T C Hooge
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, The Netherlands
| | - Diederick C Niehorster
- Lund University Humanities Lab and Department of Psychology, Lund University, Lund, Sweden
| | - Robert G Alexander
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | | | - Jeroen S Benjamins
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, The Netherlands
- Social, Health and Organizational Psychology, Utrecht University, Utrecht, The Netherlands
| | - Pieter Blignaut
- Department of Computer Science and Informatics, University of the Free State, Bloemfontein, South Africa
| | | | - Lewis L Chuang
- Department of Ergonomics, Leibniz Institute for Working Environments and Human Factors, Dortmund, Germany
- Institute of Informatics, LMU Munich, Munich, Germany
| | | | - Denis Drieghe
- School of Psychology, University of Southampton, Southampton, UK
| | - Matt J Dunn
- School of Optometry and Vision Sciences, Cardiff University, Cardiff, UK
| | | | - Susann Fiedler
- Vienna University of Economics and Business, Vienna, Austria
| | - Tom Foulsham
- Department of Psychology, University of Essex, Essex, UK
| | | | - Dan Witzner Hansen
- Machine Learning Group, Department of Computer Science, IT University of Copenhagen, Copenhagen, Denmark
| | | | - Enkelejda Kasneci
- Human-Computer Interaction, University of Tübingen, Tübingen, Germany
| | | | - Paul C Knox
- Department of Eye and Vision Science, Institute of Life Course and Medical Sciences, University of Liverpool, Liverpool, UK
| | - Ellen M Kok
- Department of Education and Pedagogy, Division Education, Faculty of Social and Behavioral Sciences, Utrecht University, Utrecht, The Netherlands
- Department of Online Learning and Instruction, Faculty of Educational Sciences, Open University of the Netherlands, Heerlen, The Netherlands
| | - Helena Lee
- University of Southampton, Southampton, UK
| | - Joy Yeonjoo Lee
- School of Health Professions Education, Faculty of Health, Medicine, and Life Sciences, Maastricht University, Maastricht, The Netherlands
| | - Jukka M Leppänen
- Department of Psychology and Speed-Language Pathology, University of Turku, Turku, Finland
| | - Stephen Macknik
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Päivi Majaranta
- TAUCHI Research Center, Computing Sciences, Faculty of Information Technology and Communication Sciences, Tampere University, Tampere, Finland
| | - Susana Martinez-Conde
- Department of Ophthalmology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Antje Nuthmann
- Institute of Psychology, University of Kiel, Kiel, Germany
| | - Marcus Nyström
- Lund University Humanities Lab, Lund University, Lund, Sweden
| | - Jacob L Orquin
- Department of Management, Aarhus University, Aarhus, Denmark
- Center for Research in Marketing and Consumer Psychology, Reykjavik University, Reykjavik, Iceland
| | - Jorge Otero-Millan
- Herbert Wertheim School of Optometry and Vision Science, University of California, Berkeley, CA, USA
| | - Soon Young Park
- Comparative Cognition, Messerli Research Institute, University of Veterinary Medicine Vienna, Medical University of Vienna, Vienna, Austria
| | - Stanislav Popelka
- Department of Geoinformatics, Palacký University Olomouc, Olomouc, Czech Republic
| | - Frank Proudlock
- The University of Leicester Ulverscroft Eye Unit, Department of Neuroscience, Psychology and Behaviour, University of Leicester, Leicester, UK
| | - Frank Renkewitz
- Department of Psychology, University of Erfurt, Erfurt, Germany
| | - Austin Roorda
- Herbert Wertheim School of Optometry and Vision Science, University of California, Berkeley, CA, USA
| | | | - Bonita Sharif
- School of Computing, University of Nebraska-Lincoln, Lincoln, Nebraska, USA
| | - Frederick Shic
- Center for Child Health, Behavior and Development, Seattle Children's Research Institute, Seattle, WA, USA
- Department of General Pediatrics, University of Washington School of Medicine, Seattle, WA, USA
| | - Mark Shovman
- Eyeviation Systems, Herzliya, Israel
- Department of Industrial Design, Bezalel Academy of Arts and Design, Jerusalem, Israel
| | - Mervyn G Thomas
- The University of Leicester Ulverscroft Eye Unit, Department of Neuroscience, Psychology and Behaviour, University of Leicester, Leicester, UK
| | - Ward Venrooij
- Electrical Engineering, Mathematics and Computer Science (EEMCS), University of Twente, Enschede, The Netherlands
| | | | - Roy S Hessels
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, The Netherlands
| |
Collapse
|
8
|
Alexander RG, Macknik SL, Martinez-Conde S. What the Neuroscience and Psychology of Magic Reveal about Misinformation. PUBLICATIONS 2022; 10:33. [PMID: 36275197 PMCID: PMC9583043 DOI: 10.3390/publications10040033] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/07/2024] Open
Abstract
When we believe misinformation, we have succumbed to an illusion: our perception or interpretation of the world does not match reality. We often trust misinformation for reasons that are unrelated to an objective, critical interpretation of the available data: Key facts go unnoticed or unreported. Overwhelming information prevents the formulation of alternative explanations. Statements become more believable every time they are repeated. Events are reframed or given "spin" to mislead audiences. In magic shows, illusionists apply similar techniques to convince spectators that false and even seemingly impossible events have happened. Yet, many magicians are "honest liars," asking audiences to suspend their disbelief only during the performance, for the sole purpose of entertainment. Magic misdirection has been studied in the lab for over a century. Psychological research has sought to understand magic from a scientific perspective and to apply the tools of magic to the understanding of cognitive and perceptual processes. More recently, neuroscientific investigations have also explored the relationship between magic illusions and their underlying brain mechanisms. We propose that the insights gained from such studies can be applied to understanding the prevalence and success of misinformation. Here, we review some of the common factors in how people experience magic during a performance and are subject to misinformation in their daily lives. Considering these factors will be important in reducing misinformation and encouraging critical thinking in society.
Collapse
Affiliation(s)
- Robert G. Alexander
- Departments of Ophthalmology, Neurology, and Physiology & Pharmacology, State University of New York Downstate Health Sciences University, Brooklyn, NY 11203, USA
| | - Stephen L. Macknik
- Departments of Ophthalmology, Neurology, and Physiology & Pharmacology, State University of New York Downstate Health Sciences University, Brooklyn, NY 11203, USA
| | - Susana Martinez-Conde
- Departments of Ophthalmology, Neurology, and Physiology & Pharmacology, State University of New York Downstate Health Sciences University, Brooklyn, NY 11203, USA
| |
Collapse
|
9
|
Joslyn S, Alexander K. Evaluating artificial intelligence algorithms for use in veterinary radiology. Vet Radiol Ultrasound 2022; 63 Suppl 1:871-879. [PMID: 36514228 DOI: 10.1111/vru.13159] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2021] [Revised: 02/16/2022] [Accepted: 03/30/2022] [Indexed: 12/15/2022] Open
Abstract
Artificial intelligence is increasingly being used for applications in veterinary radiology, including detection of abnormalities and automated measurements. Unlike human radiology, there is no formal regulation or validation of AI algorithms for veterinary medicine and both general practitioner and specialist veterinarians must rely on their own judgment when deciding whether or not to incorporate AI algorithms to aid their clinical decision-making. The benefits and challenges to developing clinically useful and diagnostically accurate AI algorithms are discussed. Considerations for the development of AI research projects are also addressed. A framework is suggested to help veterinarians, in both research and clinical practice contexts, assess AI algorithms for veterinary radiology.
Collapse
Affiliation(s)
- Steve Joslyn
- ACVR/ECVDI AI Education and Development Committee, Vedi, Perth, Western Australia, Australia
| | - Kate Alexander
- ACVR/ECVDI AI Education and Development Committee, DMV Veterinary Center, Lachine, Quebec, Canada
| |
Collapse
|
10
|
Wahid KA, Xu J, El-Habashy D, Khamis Y, Abobakr M, McDonald B, O’ Connell N, Thill D, Ahmed S, Sharafi CS, Preston K, Salzillo TC, Mohamed ASR, He R, Cho N, Christodouleas J, Fuller CD, Naser MA. Deep-learning-based generation of synthetic 6-minute MRI from 2-minute MRI for use in head and neck cancer radiotherapy. Front Oncol 2022; 12:975902. [PMID: 36425548 PMCID: PMC9679225 DOI: 10.3389/fonc.2022.975902] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2022] [Accepted: 10/21/2022] [Indexed: 11/10/2022] Open
Abstract
Background Quick magnetic resonance imaging (MRI) scans with low contrast-to-noise ratio are typically acquired for daily MRI-guided radiotherapy setup. However, for patients with head and neck (HN) cancer, these images are often insufficient for discriminating target volumes and organs at risk (OARs). In this study, we investigated a deep learning (DL) approach to generate high-quality synthetic images from low-quality images. Methods We used 108 unique HN image sets of paired 2-minute T2-weighted scans (2mMRI) and 6-minute T2-weighted scans (6mMRI). 90 image sets (~20,000 slices) were used to train a 2-dimensional generative adversarial DL model that utilized 2mMRI as input and 6mMRI as output. Eighteen image sets were used to test model performance. Similarity metrics, including the mean squared error (MSE), structural similarity index (SSIM), and peak signal-to-noise ratio (PSNR) were calculated between normalized synthetic 6mMRI and ground-truth 6mMRI for all test cases. In addition, a previously trained OAR DL auto-segmentation model was used to segment the right parotid gland, left parotid gland, and mandible on all test case images. Dice similarity coefficients (DSC) were calculated between 2mMRI and either ground-truth 6mMRI or synthetic 6mMRI for each OAR; two one-sided t-tests were applied between the ground-truth and synthetic 6mMRI to determine equivalence. Finally, a visual Turing test using paired ground-truth and synthetic 6mMRI was performed using three clinician observers; the percentage of images that were correctly identified was compared to random chance using proportion equivalence tests. Results The median similarity metrics across the whole images were 0.19, 0.93, and 33.14 for MSE, SSIM, and PSNR, respectively. The median of DSCs comparing ground-truth vs. synthetic 6mMRI auto-segmented OARs were 0.86 vs. 0.85, 0.84 vs. 0.84, and 0.82 vs. 0.85 for the right parotid gland, left parotid gland, and mandible, respectively (equivalence p<0.05 for all OARs). The percent of images correctly identified was equivalent to chance (p<0.05 for all observers). Conclusions Using 2mMRI inputs, we demonstrate that DL-generated synthetic 6mMRI outputs have high similarity to ground-truth 6mMRI, but further improvements can be made. Our study facilitates the clinical incorporation of synthetic MRI in MRI-guided radiotherapy.
Collapse
Affiliation(s)
- Kareem A. Wahid
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | | | - Dina El-Habashy
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
- Department of Clinical Oncology and Nuclear Medicine, Menoufia University, Shebin Elkom, Egypt
| | - Yomna Khamis
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
- Department of Clinical Oncology and Nuclear Medicine, Faculty of Medicine, Alexandria University, Alexandria, Egypt
| | - Moamen Abobakr
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Brigid McDonald
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | | | | | - Sara Ahmed
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Christina Setareh Sharafi
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Kathryn Preston
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Travis C. Salzillo
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Abdallah S. R. Mohamed
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Renjie He
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | | | | | - Clifton D. Fuller
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Mohamed A. Naser
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| |
Collapse
|
11
|
Phelps AM, Alexander RG, Schmidt J. Negative cues minimize visual search specificity effects. Vision Res 2022; 196:108030. [PMID: 35313163 PMCID: PMC9090971 DOI: 10.1016/j.visres.2022.108030] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/09/2021] [Revised: 03/02/2022] [Accepted: 03/08/2022] [Indexed: 11/28/2022]
Abstract
Prior target knowledge (i.e., positive cues) improves visual search performance. However, there is considerable debate about whether distractor knowledge (i.e., negative cues) can guide search. Some studies suggest the active suppression of negatively cued search items, while others suggest the initial capture of attention by negatively cued items. Prior work has used pictorial or specific text cues but has not explicitly compared them. We build on that work by comparing positive and negative cues presented pictorially and as categorical text labels using photorealistic objects and eye movement measures. Search displays contained a target (cued on positive trials), a lure from the target category (cued on negative trials), and four categorically-unrelated distractors. Search performance with positive cues resulted in stronger attentional guidance and faster object recognition for pictorial relative to categorical cues (i.e., a pictorial advantage, suggesting specific visual details afforded by pictorial cues improved search). However, in most search performance metrics, negative cues mitigate the pictorial advantage. Given that the negatively cued items captured attention, generated target guidance but mitigated the pictorial advantage, these results are partly consistent with both existing theories. Specific visual details provided in positive cues produce a large pictorial advantage in all measures, whereas specific visual details in negative cues only produce a small pictorial advantage for object recognition but not for attentional guidance. This asymmetry in the pictorial advantage suggests that the down-weighting of specific negatively cued visual features is less efficient than the up-weighting of specific positively cued visual features.
Collapse
Affiliation(s)
- Ashley M Phelps
- Department of Psychology, University of Central Florida, Orlando, FL, USA
| | - Robert G Alexander
- Departments of Ophthalmology, Neurology, and Physiology & Pharmacology, SUNY Downstate Health Sciences University, Brooklyn, NY, USA
| | - Joseph Schmidt
- Department of Psychology, University of Central Florida, Orlando, FL, USA.
| |
Collapse
|
12
|
Alexander R, Waite S, Bruno MA, Krupinski EA, Berlin L, Macknik S, Martinez-Conde S. Mandating Limits on Workload, Duty, and Speed in Radiology. Radiology 2022; 304:274-282. [PMID: 35699581 PMCID: PMC9340237 DOI: 10.1148/radiol.212631] [Citation(s) in RCA: 48] [Impact Index Per Article: 16.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/25/2023]
Abstract
Research has not yet quantified the effects of workload or duty hours on the accuracy of radiologists. With the exception of a brief reduction in imaging studies during the 2020 peak of the COVID-19 pandemic, the workload of radiologists in the United States has seen relentless growth in recent years. One concern is that this increased demand could lead to reduced accuracy. Behavioral studies in species ranging from insects to humans have shown that decision speed is inversely correlated to decision accuracy. A potential solution is to institute workload and duty limits to optimize radiologist performance and patient safety. The concern, however, is that any prescribed mandated limits would be arbitrary and thus no more advantageous than allowing radiologists to self-regulate. Specific studies have been proposed to determine whether limits reduce error, and if so, to provide a principled basis for such limits. This could determine the precise susceptibility of individual radiologists to medical error as a function of speed during image viewing, the maximum number of studies that could be read during a work shift, and the appropriate shift duration as a function of time of day. Before principled recommendations for restrictions are made, however, it is important to understand how radiologists function both optimally and at the margins of adequate performance. This study examines the relationship between interpretation speed and error rates in radiology, the potential influence of artificial intelligence on reading speed and error rates, and the possible outcomes of imposed limits on both caseload and duty hours. This review concludes that the scientific evidence needed to make meaningful rules is lacking and notes that regulating workloads without scientific principles can be more harmful than not regulating at all.
Collapse
Affiliation(s)
- Robert Alexander
- From the Departments of Ophthalmology (R.A., S.M., S.M.C.), Radiology (S.W.), Neurology (S.M., S.M.C.), and Physiology & Pharmacology (S.M., S.M.C.), SUNY Downstate Health Sciences University, 450 Clarkson Ave, Brooklyn, NY 11203; Department of Radiology, Penn State Milton S. Hershey Medical Center, Hershey, Pa (M.A.B.); Department of Radiology and Imaging Sciences, Emory University, Atlanta, Ga (E.A.K.); and Department of Radiology, Rush University Medical College and University of Illinois, Chicago, Ill (L.B.)
| | - Stephen Waite
- From the Departments of Ophthalmology (R.A., S.M., S.M.C.), Radiology (S.W.), Neurology (S.M., S.M.C.), and Physiology & Pharmacology (S.M., S.M.C.), SUNY Downstate Health Sciences University, 450 Clarkson Ave, Brooklyn, NY 11203; Department of Radiology, Penn State Milton S. Hershey Medical Center, Hershey, Pa (M.A.B.); Department of Radiology and Imaging Sciences, Emory University, Atlanta, Ga (E.A.K.); and Department of Radiology, Rush University Medical College and University of Illinois, Chicago, Ill (L.B.)
| | - Michael A Bruno
- From the Departments of Ophthalmology (R.A., S.M., S.M.C.), Radiology (S.W.), Neurology (S.M., S.M.C.), and Physiology & Pharmacology (S.M., S.M.C.), SUNY Downstate Health Sciences University, 450 Clarkson Ave, Brooklyn, NY 11203; Department of Radiology, Penn State Milton S. Hershey Medical Center, Hershey, Pa (M.A.B.); Department of Radiology and Imaging Sciences, Emory University, Atlanta, Ga (E.A.K.); and Department of Radiology, Rush University Medical College and University of Illinois, Chicago, Ill (L.B.)
| | - Elizabeth A Krupinski
- From the Departments of Ophthalmology (R.A., S.M., S.M.C.), Radiology (S.W.), Neurology (S.M., S.M.C.), and Physiology & Pharmacology (S.M., S.M.C.), SUNY Downstate Health Sciences University, 450 Clarkson Ave, Brooklyn, NY 11203; Department of Radiology, Penn State Milton S. Hershey Medical Center, Hershey, Pa (M.A.B.); Department of Radiology and Imaging Sciences, Emory University, Atlanta, Ga (E.A.K.); and Department of Radiology, Rush University Medical College and University of Illinois, Chicago, Ill (L.B.)
| | - Leonard Berlin
- From the Departments of Ophthalmology (R.A., S.M., S.M.C.), Radiology (S.W.), Neurology (S.M., S.M.C.), and Physiology & Pharmacology (S.M., S.M.C.), SUNY Downstate Health Sciences University, 450 Clarkson Ave, Brooklyn, NY 11203; Department of Radiology, Penn State Milton S. Hershey Medical Center, Hershey, Pa (M.A.B.); Department of Radiology and Imaging Sciences, Emory University, Atlanta, Ga (E.A.K.); and Department of Radiology, Rush University Medical College and University of Illinois, Chicago, Ill (L.B.)
| | - Stephen Macknik
- From the Departments of Ophthalmology (R.A., S.M., S.M.C.), Radiology (S.W.), Neurology (S.M., S.M.C.), and Physiology & Pharmacology (S.M., S.M.C.), SUNY Downstate Health Sciences University, 450 Clarkson Ave, Brooklyn, NY 11203; Department of Radiology, Penn State Milton S. Hershey Medical Center, Hershey, Pa (M.A.B.); Department of Radiology and Imaging Sciences, Emory University, Atlanta, Ga (E.A.K.); and Department of Radiology, Rush University Medical College and University of Illinois, Chicago, Ill (L.B.)
| | - Susana Martinez-Conde
- From the Departments of Ophthalmology (R.A., S.M., S.M.C.), Radiology (S.W.), Neurology (S.M., S.M.C.), and Physiology & Pharmacology (S.M., S.M.C.), SUNY Downstate Health Sciences University, 450 Clarkson Ave, Brooklyn, NY 11203; Department of Radiology, Penn State Milton S. Hershey Medical Center, Hershey, Pa (M.A.B.); Department of Radiology and Imaging Sciences, Emory University, Atlanta, Ga (E.A.K.); and Department of Radiology, Rush University Medical College and University of Illinois, Chicago, Ill (L.B.)
| |
Collapse
|
13
|
He Q, Yang XY, Zhao D, Fang F. Enhancement of visual perception by combining transcranial electrical stimulation and visual perceptual training. MEDICAL REVIEW (BERLIN, GERMANY) 2022; 2:271-284. [PMID: 37724187 PMCID: PMC10388778 DOI: 10.1515/mr-2022-0010] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/19/2022] [Accepted: 06/16/2022] [Indexed: 09/20/2023]
Abstract
The visual system remains highly malleable even after its maturity or impairment. Our visual function can be enhanced through many ways, such as transcranial electrical stimulation (tES) and visual perceptual learning (VPL). TES can change visual function rapidly, but its modulation effect is short-lived and unstable. By contrast, VPL can lead to a substantial and long-lasting improvement in visual function, but extensive training is typically required. Theoretically, visual function could be further improved in a shorter time frame by combining tES and VPL than by solely using tES or VPL. Vision enhancement by combining these two methods concurrently is both theoretically and practically significant. In this review, we firstly introduced the basic concept and possible mechanisms of VPL and tES; then we reviewed the current research progress of visual enhancement using the combination of two methods in both general and clinical population; finally, we discussed the limitations and future directions in this field. Our review provides a guide for future research and application of vision enhancement and restoration by combining VPL and tES.
Collapse
Affiliation(s)
- Qing He
- School of Psychological and Cognitive Sciences and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing, China
- Key Laboratory of Machine Perception, Ministry of Education, Peking University, Beijing, China
- IDG/McGovern Institute for Brain Research, Peking University, Beijing, China
- Peking-Tsinghua Center for Life Sciences, Peking University, Beijing, China
| | - Xin-Yue Yang
- School of Psychological and Cognitive Sciences and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing, China
- Key Laboratory of Machine Perception, Ministry of Education, Peking University, Beijing, China
- IDG/McGovern Institute for Brain Research, Peking University, Beijing, China
- Peking-Tsinghua Center for Life Sciences, Peking University, Beijing, China
| | - Daiqing Zhao
- Department of Psychology, The Pennsylvania State University, University Park, State College, PA, USA
| | - Fang Fang
- School of Psychological and Cognitive Sciences and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing, China
- Key Laboratory of Machine Perception, Ministry of Education, Peking University, Beijing, China
- IDG/McGovern Institute for Brain Research, Peking University, Beijing, China
- Peking-Tsinghua Center for Life Sciences, Peking University, Beijing, China
| |
Collapse
|
14
|
Alexander RG, Yazdanie F, Waite S, Chaudhry ZA, Kolla S, Macknik SL, Martinez-Conde S. Visual Illusions in Radiology: Untrue Perceptions in Medical Images and Their Implications for Diagnostic Accuracy. Front Neurosci 2021; 15:629469. [PMID: 34177444 PMCID: PMC8226024 DOI: 10.3389/fnins.2021.629469] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/17/2020] [Accepted: 04/19/2021] [Indexed: 11/13/2022] Open
Abstract
Errors in radiologic interpretation are largely the result of failures of perception. This remains true despite the increasing use of computer-aided detection and diagnosis. We surveyed the literature on visual illusions during the viewing of radiologic images. Misperception of anatomical structures is a potential cause of error that can lead to patient harm if disease is seen when none is present. However, visual illusions can also help enhance the ability of radiologists to detect and characterize abnormalities. Indeed, radiologists have learned to exploit certain perceptual biases in diagnostic findings and as training tools. We propose that further detailed study of radiologic illusions would help clarify the mechanisms underlying radiologic performance and provide additional heuristics to improve radiologist training and reduce medical error.
Collapse
Affiliation(s)
- Robert G Alexander
- Department of Ophthalmology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States.,Department of Neurology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States.,Department of Physiology and Pharmacology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| | - Fahd Yazdanie
- Department of Radiology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| | - Stephen Waite
- Department of Radiology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| | - Zeshan A Chaudhry
- Department of Radiology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| | - Srinivas Kolla
- Department of Radiology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| | - Stephen L Macknik
- Department of Ophthalmology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States.,Department of Neurology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States.,Department of Physiology and Pharmacology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| | - Susana Martinez-Conde
- Department of Ophthalmology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States.,Department of Neurology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States.,Department of Physiology and Pharmacology, State University of New York Downstate Health Sciences University, Brooklyn, NY, United States
| |
Collapse
|