1
|
Kim MS, Park B, Sippel GJ, Mun AH, Yang W, McCarthy KH, Fernandez E, Linguraru MG, Sarcevic A, Marsic I, Burd RS. Comparative analysis of personal protective equipment nonadherence detection: computer vision versus human observers. J Am Med Inform Assoc 2025; 32:163-171. [PMID: 39401253 DOI: 10.1093/jamia/ocae262] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/19/2024] [Revised: 09/03/2024] [Accepted: 10/01/2024] [Indexed: 12/15/2024] Open
Abstract
OBJECTIVES Human monitoring of personal protective equipment (PPE) adherence among healthcare providers has several limitations, including the need for additional personnel during staff shortages and decreased vigilance during prolonged tasks. To address these challenges, we developed an automated computer vision system for monitoring PPE adherence in healthcare settings. We assessed the system performance against human observers detecting nonadherence in a video surveillance experiment. MATERIALS AND METHODS The automated system was trained to detect 15 classes of eyewear, masks, gloves, and gowns using an object detector and tracker. To assess how the system performs compared to human observers in detecting nonadherence, we designed a video surveillance experiment under 2 conditions: variations in video durations (20, 40, and 60 seconds) and the number of individuals in the videos (3 versus 6). Twelve nurses participated as human observers. Performance was assessed based on the number of detections of nonadherence. RESULTS Human observers detected fewer instances of nonadherence than the system (parameter estimate -0.3, 95% CI -0.4 to -0.2, P < .001). Human observers detected more nonadherence during longer video durations (parameter estimate 0.7, 95% CI 0.4-1.0, P < .001). The system achieved a sensitivity of 0.86, specificity of 1, and Matthew's correlation coefficient of 0.82 for detecting PPE nonadherence. DISCUSSION An automated system simultaneously tracks multiple objects and individuals. The system performance is also independent of observation duration, an improvement over human monitoring. CONCLUSION The automated system presents a potential solution for scalable monitoring of hospital-wide infection control practices and improving PPE usage in healthcare settings.
Collapse
Affiliation(s)
- Mary S Kim
- Division of Trauma and Burn Surgery, Children's National Hospital, Washington, DC 20010, United States
| | - Beomseok Park
- Department of Electrical and Computer Engineering, Rutgers University, New Brunswick, NJ 08901, United States
| | - Genevieve J Sippel
- Division of Trauma and Burn Surgery, Children's National Hospital, Washington, DC 20010, United States
| | - Aaron H Mun
- Division of Trauma and Burn Surgery, Children's National Hospital, Washington, DC 20010, United States
| | - Wanzhao Yang
- Department of Electrical and Computer Engineering, Rutgers University, New Brunswick, NJ 08901, United States
| | - Kathleen H McCarthy
- Division of Trauma and Burn Surgery, Children's National Hospital, Washington, DC 20010, United States
| | - Emely Fernandez
- Division of Trauma and Burn Surgery, Children's National Hospital, Washington, DC 20010, United States
| | - Marius George Linguraru
- Sheikh Zayed Institute for Pediatric Surgical Innovation, Children's National Hospital, Washington, DC 20010, United States
- Departments of Radiology and Pediatrics, George Washington University, Washington, DC 20037, United States
| | - Aleksandra Sarcevic
- College of Computing and Informatics, Drexel University, Philadelphia, PA 19104, United States
| | - Ivan Marsic
- Department of Electrical and Computer Engineering, Rutgers University, New Brunswick, NJ 08901, United States
| | - Randall S Burd
- Division of Trauma and Burn Surgery, Children's National Hospital, Washington, DC 20010, United States
| |
Collapse
|
2
|
Lawrence P, Brown C. Comparing Human-Level and Machine Learning Model Performance in White Blood Cell Morphology Assessment. Eur J Haematol 2024. [PMID: 39370635 DOI: 10.1111/ejh.14318] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/26/2024] [Revised: 09/24/2024] [Accepted: 09/25/2024] [Indexed: 10/08/2024]
Abstract
INTRODUCTION There is an increasing research focus on the role of machine learning in the haematology laboratory, particularly in blood cell morphologic assessment. Human-level performance is an important baseline and goal for machine learning. This study aims to assess the interobserver variability and human-level performance in blood cell morphologic assessment. METHODS A dataset of 1000 single white blood cell images were independently labelled by 10 doctors and morphology scientists. Interobserver variability was calculated using Fleiss' kappa. Observers' labels were then separated into consensus labels used to determine ground truth, and performance labels used to assess observer performance. A machine learning model was trained and assessed using the same cell images. Explainability images (XRAI and IG) were generated for each of the test images. RESULTS The Fleiss kappa for all 10 observers was 0.608, indicating substantial agreement between observers. The accuracy of human observers was 95%, with sensitivity 72% and specificity 97%. The accuracy of the machine learning model was 95%, with sensitivity 71% and specificity 97%. The model shared similar performance across labels when compared to humans. Explainability metrics demonstrated that the machine learning model was able to differentiate between the cytoplasm and nucleus of the cells, and used these features to perform predictions. CONCLUSION The substantial, though not perfect, agreement between human observers highlights the inherent subjectivity in white blood cell morphologic assessment. A machine learning model performed similarly to human observers in single white blood cell identification. Further research is needed to compare human-level and machine learning performance in ways that more closely reflect the typical process of morphologic assessment.
Collapse
|
3
|
Chaibutr N, Pongpanitanont P, Laymanivong S, Thanchomnang T, Janwan P. Development of a Machine Learning Model for the Classification of Enterobius vermicularis Egg. J Imaging 2024; 10:212. [PMID: 39330432 PMCID: PMC11433018 DOI: 10.3390/jimaging10090212] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2024] [Revised: 08/22/2024] [Accepted: 08/24/2024] [Indexed: 09/28/2024] Open
Abstract
Enterobius vermicularis (pinworm) infections are a significant global health issue, affecting children predominantly in environments like schools and daycares. Traditional diagnosis using the scotch tape technique involves examining E. vermicularis eggs under a microscope. This method is time-consuming and depends heavily on the examiner's expertise. To improve this, convolutional neural networks (CNNs) have been used to automate the detection of pinworm eggs from microscopic images. In our study, we enhanced E. vermicularis egg detection using a CNN benchmarked against leading models. We digitized and augmented 40,000 images of E. vermicularis eggs (class 1) and artifacts (class 0) for comprehensive training, using an 80:20 training-validation and a five-fold cross-validation. The proposed CNN model showed limited initial performance but achieved 90.0% accuracy, precision, recall, and F1-score after data augmentation. It also demonstrated improved stability with an ROC-AUC metric increase from 0.77 to 0.97. Despite its smaller file size, our CNN model performed comparably to larger models. Notably, the Xception model achieved 99.0% accuracy, precision, recall, and F1-score. These findings highlight the effectiveness of data augmentation and advanced CNN architectures in improving diagnostic accuracy and efficiency for E. vermicularis infections.
Collapse
Affiliation(s)
- Natthanai Chaibutr
- Medical Innovation and Technology Program, School of Allied Health Sciences, Walailak University, Nakhon Si Thammarat 80160, Thailand
- Hematology and Transfusion Science Research Center, Walailak University, Nakhon Si Thammarat 80160, Thailand
- Medical Technology Service Center, Prince of Songkla University, Phuket Campus, Phuket 83120, Thailand
| | - Pongphan Pongpanitanont
- Health Sciences (International Program), College of Graduate Studies, Walailak University, Nakhon Si Thammarat 80160, Thailand
| | - Sakhone Laymanivong
- Centre of Malariology, Parasitology and Entomology, Ministry of Health, Vientiane Capital P.O. Box 0100, Laos
| | | | - Penchom Janwan
- Medical Innovation and Technology Program, School of Allied Health Sciences, Walailak University, Nakhon Si Thammarat 80160, Thailand
- Hematology and Transfusion Science Research Center, Walailak University, Nakhon Si Thammarat 80160, Thailand
- Department of Medical Technology, School of Allied Health Sciences, Walailak University, Nakhon Si Thammarat 80160, Thailand
| |
Collapse
|
4
|
Hosoda K, Nishida K, Seno S, Mashita T, Kashioka H, Ohzawa I. A single fast Hebbian-like process enabling one-shot class addition in deep neural networks without backbone modification. Front Neurosci 2024; 18:1344114. [PMID: 38933813 PMCID: PMC11202076 DOI: 10.3389/fnins.2024.1344114] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2023] [Accepted: 05/16/2024] [Indexed: 06/28/2024] Open
Abstract
One-shot learning, the ability to learn a new concept from a single instance, is a distinctive brain function that has garnered substantial interest in machine learning. While modeling physiological mechanisms poses challenges, advancements in artificial neural networks have led to performances in specific tasks that rival human capabilities. Proposing one-shot learning methods with these advancements, especially those involving simple mechanisms, not only enhance technological development but also contribute to neuroscience by proposing functionally valid hypotheses. Among the simplest methods for one-shot class addition with deep learning image classifiers is "weight imprinting," which uses neural activity from a new class image data as the corresponding new synaptic weights. Despite its simplicity, its relevance to neuroscience is ambiguous, and it often interferes with original image classification, which is a significant drawback in practical applications. This study introduces a novel interpretation where a part of the weight imprinting process aligns with the Hebbian rule. We show that a single Hebbian-like process enables pre-trained deep learning image classifiers to perform one-shot class addition without any modification to the original classifier's backbone. Using non-parametric normalization to mimic brain's fast Hebbian plasticity significantly reduces the interference observed in previous methods. Our method is one of the simplest and most practical for one-shot class addition tasks, and its reliance on a single fast Hebbian-like process contributes valuable insights to neuroscience hypotheses.
Collapse
Affiliation(s)
- Kazufumi Hosoda
- Center for Information and Neural Networks, Advanced ICT Research Institute, National Institute of Information and Communications Technology, Suita, Japan
- Life and Medical Sciences Area, Health Sciences Discipline, Kobe University, Kobe, Japan
| | - Keigo Nishida
- Laboratory for Computational Molecular Design, RIKEN Center for Biosystems Dynamics Research, Suita, Japan
| | - Shigeto Seno
- Department of Bioinformatic Engineering, Graduate School of Information Science and Technology, Osaka University, Suita, Japan
| | | | - Hideki Kashioka
- Center for Information and Neural Networks, Advanced ICT Research Institute, National Institute of Information and Communications Technology, Suita, Japan
| | - Izumi Ohzawa
- Center for Information and Neural Networks, Advanced ICT Research Institute, National Institute of Information and Communications Technology, Suita, Japan
| |
Collapse
|
5
|
Perception without preconception: comparison between the human and machine learner in recognition of tissues from histological sections. Sci Rep 2022; 12:16420. [PMID: 36180472 PMCID: PMC9525725 DOI: 10.1038/s41598-022-20012-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/27/2021] [Accepted: 09/07/2022] [Indexed: 11/08/2022] Open
Abstract
Deep neural networks (DNNs) have shown success in image classification, with high accuracy in recognition of everyday objects. Performance of DNNs has traditionally been measured assuming human accuracy is perfect. In specific problem domains, however, human accuracy is less than perfect and a comparison between humans and machine learning (ML) models can be performed. In recognising everyday objects, humans have the advantage of a lifetime of experience, whereas DNN models are trained only with a limited image dataset. We have tried to compare performance of human learners and two DNN models on an image dataset which is novel to both, i.e. histological images. We thus aim to eliminate the advantage of prior experience that humans have over DNN models in image classification. Ten classes of tissues were randomly selected from the undergraduate first year histology curriculum of a Medical School in North India. Two machine learning (ML) models were developed based on the VGG16 (VML) and Inception V2 (IML) DNNs, using transfer learning, to produce a 10-class classifier. One thousand (1000) images belonging to the ten classes (i.e. 100 images from each class) were split into training (700) and validation (300) sets. After training, the VML and IML model achieved 85.67 and 89% accuracy on the validation set, respectively. The training set was also circulated to medical students (MS) of the college for a week. An online quiz, consisting of a random selection of 100 images from the validation set, was conducted on students (after obtaining informed consent) who volunteered for the study. 66 students participated in the quiz, providing 6557 responses. In addition, we prepared a set of 10 images which belonged to different classes of tissue, not present in training set (i.e. out of training scope or OTS images). A second quiz was conducted on medical students with OTS images, and the ML models were also run on these OTS images. The overall accuracy of MS in the first quiz was 55.14%. The two ML models were also run on the first quiz questionnaire, producing accuracy between 91 and 93%. The ML models scored more than 80% of medical students. Analysis of confusion matrices of both ML models and all medical students showed dissimilar error profiles. However, when comparing the subset of students who achieved similar accuracy as the ML models, the error profile was also similar. Recognition of 'stomach' proved difficult for both humans and ML models. In 04 images in the first quiz set, both VML model and medical students produced highly equivocal responses. Within these images, a pattern of bias was uncovered-the tendency of medical students to misclassify 'liver' tissue. The 'stomach' class proved most difficult for both MS and VML, producing 34.84% of all errors of MS, and 41.17% of all errors of VML model; however, the IML model committed most errors in recognising the 'skin' class (27.5% of all errors). Analysis of the convolution layers of the DNN outlined features in the original image which might have led to misclassification by the VML model. In OTS images, however, the medical students produced better overall score than both ML models, i.e. they successfully recognised patterns of similarity between tissues and could generalise their training to a novel dataset. Our findings suggest that within the scope of training, ML models perform better than 80% medical students with a distinct error profile. However, students who have reached accuracy close to the ML models, tend to replicate the error profile as that of the ML models. This suggests a degree of similarity between how machines and humans extract features from an image. If asked to recognise images outside the scope of training, humans perform better at recognising patterns and likeness between tissues. This suggests that 'training' is not the same as 'learning', and humans can extend their pattern-based learning to different domains outside of the training set.
Collapse
|