1
|
López-Pérez M, Morquecho A, Schmidt A, Pérez-Bueno F, Martín-Castro A, Mateos J, Molina R. The CrowdGleason dataset: Learning the Gleason grade from crowds and experts. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 257:108472. [PMID: 39488043 DOI: 10.1016/j.cmpb.2024.108472] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/02/2024] [Revised: 09/30/2024] [Accepted: 10/20/2024] [Indexed: 11/04/2024]
Abstract
BACKGROUND Currently, prostate cancer (PCa) diagnosis relies on the human analysis of prostate biopsy Whole Slide Images (WSIs) using the Gleason score. Since this process is error-prone and time-consuming, recent advances in machine learning have promoted the use of automated systems to assist pathologists. Unfortunately, labeled datasets for training and validation are scarce due to the need for expert pathologists to provide ground-truth labels. METHODS This work introduces a new prostate histopathological dataset named CrowdGleason, which consists of 19,077 patches from 1045 WSIs with various Gleason grades. The dataset was annotated using a crowdsourcing protocol involving seven pathologists-in-training to distribute the labeling effort. To provide a baseline analysis, two crowdsourcing methods based on Gaussian Processes (GPs) were evaluated for Gleason grade prediction: SVGPCR, which learns a model from the CrowdGleason dataset, and SVGPMIX, which combines data from the public dataset SICAPv2 and the CrowdGleason dataset. The performance of these methods was compared with other crowdsourcing and expert label-based methods through comprehensive experiments. RESULTS The results demonstrate that our GP-based crowdsourcing approach outperforms other methods for aggregating crowdsourced labels (κ=0.7048±0.0207) for SVGPCR vs.(κ=0.6576±0.0086) for SVGP with majority voting). SVGPCR trained with crowdsourced labels performs better than GP trained with expert labels from SICAPv2 (κ=0.6583±0.0220) and outperforms most individual pathologists-in-training (mean κ=0.5432). Additionally, SVGPMIX trained with a combination of SICAPv2 and CrowdGleason achieves the highest performance on both datasets (κ=0.7814±0.0083 and κ=0.7276±0.0260). CONCLUSION The experiments show that the CrowdGleason dataset can be successfully used for training and validating supervised and crowdsourcing methods. Furthermore, the crowdsourcing methods trained on this dataset obtain competitive results against those using expert labels. Interestingly, the combination of expert and non-expert labels opens the door to a future of massive labeling by incorporating both expert and non-expert pathologist annotators.
Collapse
Affiliation(s)
- Miguel López-Pérez
- Instituto Universitario de Investigación en Tecnología Centrada en el Ser Humano, Universitat Politècnica de València, Spain.
| | - Alba Morquecho
- Department of Computer Science and Artificial Intelligence, Universidad de Granada, Granada, Spain.
| | - Arne Schmidt
- Department of Computer Science and Artificial Intelligence, Universidad de Granada, Granada, Spain.
| | - Fernando Pérez-Bueno
- Basque Center on Cognition, Brain and Language, Donostia - San Sebastián, Spain.
| | - Aurelio Martín-Castro
- Department of Pathology, Virgen de las Nieves University Hospital, 18014 Granada, Spain.
| | - Javier Mateos
- Department of Computer Science and Artificial Intelligence, Universidad de Granada, Granada, Spain.
| | - Rafael Molina
- Department of Computer Science and Artificial Intelligence, Universidad de Granada, Granada, Spain.
| |
Collapse
|
2
|
Seoni S, Shahini A, Meiburger KM, Marzola F, Rotunno G, Acharya UR, Molinari F, Salvi M. All you need is data preparation: A systematic review of image harmonization techniques in Multi-center/device studies for medical support systems. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 250:108200. [PMID: 38677080 DOI: 10.1016/j.cmpb.2024.108200] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/27/2024] [Revised: 04/20/2024] [Accepted: 04/22/2024] [Indexed: 04/29/2024]
Abstract
BACKGROUND AND OBJECTIVES Artificial intelligence (AI) models trained on multi-centric and multi-device studies can provide more robust insights and research findings compared to single-center studies. However, variability in acquisition protocols and equipment can introduce inconsistencies that hamper the effective pooling of multi-source datasets. This systematic review evaluates strategies for image harmonization, which standardizes appearances to enable reliable AI analysis of multi-source medical imaging. METHODS A literature search using PRISMA guidelines was conducted to identify relevant papers published between 2013 and 2023 analyzing multi-centric and multi-device medical imaging studies that utilized image harmonization approaches. RESULTS Common image harmonization techniques included grayscale normalization (improving classification accuracy by up to 24.42 %), resampling (increasing the percentage of robust radiomics features from 59.5 % to 89.25 %), and color normalization (enhancing AUC by up to 0.25 in external test sets). Initially, mathematical and statistical methods dominated, but machine and deep learning adoption has risen recently. Color imaging modalities like digital pathology and dermatology have remained prominent application areas, though harmonization efforts have expanded to diverse fields including radiology, nuclear medicine, and ultrasound imaging. In all the modalities covered by this review, image harmonization improved AI performance, with increasing of up to 24.42 % in classification accuracy and 47 % in segmentation Dice scores. CONCLUSIONS Continued progress in image harmonization represents a promising strategy for advancing healthcare by enabling large-scale, reliable analysis of integrated multi-source datasets using AI. Standardizing imaging data across clinical settings can help realize personalized, evidence-based care supported by data-driven technologies while mitigating biases associated with specific populations or acquisition protocols.
Collapse
Affiliation(s)
- Silvia Seoni
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy
| | - Alen Shahini
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy
| | - Kristen M Meiburger
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy
| | - Francesco Marzola
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy
| | - Giulia Rotunno
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy
| | - U Rajendra Acharya
- School of Mathematics, Physics and Computing, University of Southern Queensland, Springfield, Australia; Centre for Health Research, University of Southern Queensland, Australia
| | - Filippo Molinari
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy
| | - Massimo Salvi
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy.
| |
Collapse
|
3
|
McGenity C, Clarke EL, Jennings C, Matthews G, Cartlidge C, Freduah-Agyemang H, Stocken DD, Treanor D. Artificial intelligence in digital pathology: a systematic review and meta-analysis of diagnostic test accuracy. NPJ Digit Med 2024; 7:114. [PMID: 38704465 PMCID: PMC11069583 DOI: 10.1038/s41746-024-01106-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/17/2023] [Accepted: 04/12/2024] [Indexed: 05/06/2024] Open
Abstract
Ensuring diagnostic performance of artificial intelligence (AI) before introduction into clinical practice is essential. Growing numbers of studies using AI for digital pathology have been reported over recent years. The aim of this work is to examine the diagnostic accuracy of AI in digital pathology images for any disease. This systematic review and meta-analysis included diagnostic accuracy studies using any type of AI applied to whole slide images (WSIs) for any disease. The reference standard was diagnosis by histopathological assessment and/or immunohistochemistry. Searches were conducted in PubMed, EMBASE and CENTRAL in June 2022. Risk of bias and concerns of applicability were assessed using the QUADAS-2 tool. Data extraction was conducted by two investigators and meta-analysis was performed using a bivariate random effects model, with additional subgroup analyses also performed. Of 2976 identified studies, 100 were included in the review and 48 in the meta-analysis. Studies were from a range of countries, including over 152,000 whole slide images (WSIs), representing many diseases. These studies reported a mean sensitivity of 96.3% (CI 94.1-97.7) and mean specificity of 93.3% (CI 90.5-95.4). There was heterogeneity in study design and 99% of studies identified for inclusion had at least one area at high or unclear risk of bias or applicability concerns. Details on selection of cases, division of model development and validation data and raw performance data were frequently ambiguous or missing. AI is reported as having high diagnostic accuracy in the reported areas but requires more rigorous evaluation of its performance.
Collapse
Affiliation(s)
- Clare McGenity
- University of Leeds, Leeds, UK.
- Leeds Teaching Hospitals NHS Trust, Leeds, UK.
| | - Emily L Clarke
- University of Leeds, Leeds, UK
- Leeds Teaching Hospitals NHS Trust, Leeds, UK
| | - Charlotte Jennings
- University of Leeds, Leeds, UK
- Leeds Teaching Hospitals NHS Trust, Leeds, UK
| | | | | | | | | | - Darren Treanor
- University of Leeds, Leeds, UK
- Leeds Teaching Hospitals NHS Trust, Leeds, UK
- Department of Clinical Pathology and Department of Clinical and Experimental Medicine, Linköping University, Linköping, Sweden
- Centre for Medical Image Science and Visualization (CMIV), Linköping University, Linköping, Sweden
| |
Collapse
|
4
|
Kanwal N, López-Pérez M, Kiraz U, Zuiverloon TCM, Molina R, Engan K. Are you sure it's an artifact? Artifact detection and uncertainty quantification in histological images. Comput Med Imaging Graph 2024; 112:102321. [PMID: 38199127 DOI: 10.1016/j.compmedimag.2023.102321] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2023] [Revised: 11/08/2023] [Accepted: 12/12/2023] [Indexed: 01/12/2024]
Abstract
Modern cancer diagnostics involves extracting tissue specimens from suspicious areas and conducting histotechnical procedures to prepare a digitized glass slide, called Whole Slide Image (WSI), for further examination. These procedures frequently introduce different types of artifacts in the obtained WSI, and histological artifacts might influence Computational Pathology (CPATH) systems further down to a diagnostic pipeline if not excluded or handled. Deep Convolutional Neural Networks (DCNNs) have achieved promising results for the detection of some WSI artifacts, however, they do not incorporate uncertainty in their predictions. This paper proposes an uncertainty-aware Deep Kernel Learning (DKL) model to detect blurry areas and folded tissues, two types of artifacts that can appear in WSIs. The proposed probabilistic model combines a CNN feature extractor and a sparse Gaussian Processes (GPs) classifier, which improves the performance of current state-of-the-art artifact detection DCNNs and provides uncertainty estimates. We achieved 0.996 and 0.938 F1 scores for blur and folded tissue detection on unseen data, respectively. In extensive experiments, we validated the DKL model on unseen data from external independent cohorts with different staining and tissue types, where it outperformed DCNNs. Interestingly, the DKL model is more confident in the correct predictions and less in the wrong ones. The proposed DKL model can be integrated into the preprocessing pipeline of CPATH systems to provide reliable predictions and possibly serve as a quality control tool.
Collapse
Affiliation(s)
- Neel Kanwal
- Department of Electrical Engineering and Computer Science, University of Stavanger, 4021 Stavanger, Norway.
| | - Miguel López-Pérez
- Department of Computer Science and Artificial Intelligence, University of Granada, 18071 Granada, Spain
| | - Umay Kiraz
- Department of Pathology, Stavanger University Hospital, 4011 Stavanger, Norway; Department of Chemistry, Bioscience and Environmental Engineering, University of Stavanger, 4021 Stavanger, Norway
| | - Tahlita C M Zuiverloon
- Department of Urology, University Medical Center Rotterdam, Erasmus MC Cancer Institute, 1035 GD Rotterdam, The Netherlands
| | - Rafael Molina
- Department of Computer Science and Artificial Intelligence, University of Granada, 18071 Granada, Spain
| | - Kjersti Engan
- Department of Electrical Engineering and Computer Science, University of Stavanger, 4021 Stavanger, Norway
| |
Collapse
|
5
|
G K AV, Gogoi G, Kachappilly MC, Rangarajan A, Pandya HJ. Label-free multimodal electro-thermo-mechanical (ETM) phenotyping as a novel biomarker to differentiate between normal, benign, and cancerous breast biopsy tissues. J Biol Eng 2023; 17:68. [PMID: 37957665 PMCID: PMC10644568 DOI: 10.1186/s13036-023-00388-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2023] [Accepted: 10/30/2023] [Indexed: 11/15/2023] Open
Abstract
BACKGROUND Technologies for quick and label-free diagnosis of malignancies from breast tissues have the potential to be a significant adjunct to routine diagnostics. The biophysical phenotypes of breast tissues, such as its electrical, thermal, and mechanical properties (ETM), have the potential to serve as novel markers to differentiate between normal, benign, and malignant tissue. RESULTS We report a system-of-biochips (SoB) integrated into a semi-automated mechatronic system that can characterize breast biopsy tissues using electro-thermo-mechanical sensing. The SoB, fabricated on silicon using microfabrication techniques, can measure the electrical impedance (Z), thermal conductivity (K), mechanical stiffness (k), and viscoelastic stress relaxation (%R) of the samples. The key sensing elements of the biochips include interdigitated electrodes, resistance temperature detectors, microheaters, and a micromachined diaphragm with piezoresistive bridges. Multi-modal ETM measurements performed on formalin-fixed tumour and adjacent normal breast biopsy samples from N = 14 subjects were able to differentiate between invasive ductal carcinoma (malignant), fibroadenoma (benign), and adjacent normal (healthy) tissues with a root mean square error of 0.2419 using a Gaussian process classifier. Carcinoma tissues were observed to have the highest mean impedance (110018.8 ± 20293.8 Ω) and stiffness (0.076 ± 0.009 kNm-1) and the lowest thermal conductivity (0.189 ± 0.019 Wm-1 K-1) amongst the three groups, while the fibroadenoma samples had the highest percentage relaxation in normalized load (47.8 ± 5.12%). CONCLUSIONS The work presents a novel strategy to characterize the multi-modal biophysical phenotype of breast biopsy tissues to aid in cancer diagnosis from small-sized tumour samples. The methodology envisions to supplement the existing technology gap in the analysis of breast tissue samples in the pathology laboratories to aid the diagnostic workflow.
Collapse
Affiliation(s)
- Anil Vishnu G K
- Center for BioSystems Science and Engineering, Indian Institute of Science, Bangalore, Karnataka, 560012, India
| | - Gayatri Gogoi
- Department of Pathology, Assam Medical College, Dibrugarh, Assam, 786002, India
| | - Midhun C Kachappilly
- Department of Electronic Systems Engineering, Indian Institute of Science, Bangalore, Karnataka, 560012, India
| | - Annapoorni Rangarajan
- Department of Developmental Biology and Genetics, Indian Institute of Science, Bangalore, Karnataka, 560012, India
| | - Hardik J Pandya
- Department of Electronic Systems Engineering, Indian Institute of Science, Bangalore, Karnataka, 560012, India.
- Centre for Product Design and Manufacturing, Indian Institute of Science, Bangalore, Karnataka, 560012, India.
| |
Collapse
|
6
|
Parwani AV, Patel A, Zhou M, Cheville JC, Tizhoosh H, Humphrey P, Reuter VE, True LD. An update on computational pathology tools for genitourinary pathology practice: A review paper from the Genitourinary Pathology Society (GUPS). J Pathol Inform 2023; 14:100177. [PMID: 36654741 PMCID: PMC9841212 DOI: 10.1016/j.jpi.2022.100177] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2022] [Revised: 12/20/2022] [Accepted: 12/20/2022] [Indexed: 12/31/2022] Open
Abstract
Machine learning has been leveraged for image analysis applications throughout a multitude of subspecialties. This position paper provides a perspective on the evolutionary trajectory of practical deep learning tools for genitourinary pathology through evaluating the most recent iterations of such algorithmic devices. Deep learning tools for genitourinary pathology demonstrate potential to enhance prognostic and predictive capacity for tumor assessment including grading, staging, and subtype identification, yet limitations in data availability, regulation, and standardization have stymied their implementation.
Collapse
Affiliation(s)
- Anil V. Parwani
- The Ohio State University, Columbus, Ohio, USA
- Corresponding author.
| | - Ankush Patel
- The Ohio State University, 2441 60th Ave SE, Mercer Island, Washington 98040, USA
| | - Ming Zhou
- Tufts University, Medford, Massachusetts, USA
| | | | | | | | | | | |
Collapse
|
7
|
Patel AU, Mohanty SK, Parwani AV. Applications of Digital and Computational Pathology and Artificial Intelligence in Genitourinary Pathology Diagnostics. Surg Pathol Clin 2022; 15:759-785. [PMID: 36344188 DOI: 10.1016/j.path.2022.08.001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
As machine learning (ML) solutions for genitourinary pathology image analysis are fostered by a progressively digitized laboratory landscape, these integrable modalities usher in a revolution in histopathological diagnosis. As technology advances, limitations stymying clinical artificial intelligence (AI) will not be extinguished without thorough validation and interrogation of ML tools by pathologists and regulatory bodies alike. ML solutions deployed in clinical settings for applications in prostate pathology yield promising results. Recent breakthroughs in clinical artificial intelligence for genitourinary pathology demonstrate unprecedented generalizability, heralding prospects for a future in which AI-driven assistive solutions may be seen as laboratory faculty, rather than novelty.
Collapse
Affiliation(s)
- Ankush Uresh Patel
- Department of Laboratory Medicine and Pathology, Mayo Clinic, 200 First Street Southwest, Rochester, MN 55905, USA
| | - Sambit K Mohanty
- Surgical and Molecular Pathology, Advanced Medical Research Institute, Plot No. 1, Near Jayadev Vatika Park, Khandagiri, Bhubaneswar, Odisha 751019. https://twitter.com/SAMBITKMohanty1
| | - Anil V Parwani
- Department of Pathology, The Ohio State University, Cooperative Human Tissue Network (CHTN) Midwestern Division Polaris Innovation Centre, 2001 Polaris Parkway Suite 1000, Columbus, OH 43240, USA.
| |
Collapse
|
8
|
Paulson N, Zeevi T, Papademetris M, Leapman MS, Onofrey JA, Sprenkle PC, Humphrey PA, Staib LH, Levi AW. Prediction of Adverse Pathology at Radical Prostatectomy in Grade Group 2 and 3 Prostate Biopsies Using Machine Learning. JCO Clin Cancer Inform 2022; 6:e2200016. [PMID: 36179281 DOI: 10.1200/cci.22.00016] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022] Open
Abstract
PURPOSE There is ongoing clinical need to improve estimates of disease outcome in prostate cancer. Machine learning (ML) approaches to pathologic diagnosis and prognosis are a promising and increasingly used strategy. In this study, we use an ML algorithm for prediction of adverse outcomes at radical prostatectomy (RP) using whole-slide images (WSIs) of prostate biopsies with Grade Group (GG) 2 or 3 disease. METHODS We performed a retrospective review of prostate biopsies collected at our institution which had corresponding RP, GG 2 or 3 disease one or more cores, and no biopsies with higher than GG 3 disease. A hematoxylin and eosin-stained core needle biopsy from each site with GG 2 or 3 disease was scanned and used as the sole input for the algorithm. The ML pipeline had three phases: image preprocessing, feature extraction, and adverse outcome prediction. First, patches were extracted from each biopsy scan. Subsequently, the pre-trained Visual Geometry Group-16 convolutional neural network was used for feature extraction. A representative feature vector was then used as input to an Extreme Gradient Boosting classifier for predicting the binary adverse outcome. We subsequently assessed patient clinical risk using CAPRA score for comparison with the ML pipeline results. RESULTS The data set included 361 WSIs from 107 patients (56 with adverse pathology at RP). The area under the receiver operating characteristic curves for the ML classification were 0.72 (95% CI, 0.62 to 0.81), 0.65 (95% CI, 0.53 to 0.79) and 0.89 (95% CI, 0.79 to 1.00) for the entire cohort, and GG 2 and GG 3 patients, respectively, similar to the performance of the CAPRA clinical risk assessment. CONCLUSION We provide evidence for the potential of ML algorithms to use WSIs of needle core prostate biopsies to estimate clinically relevant prostate cancer outcomes.
Collapse
Affiliation(s)
| | - Tal Zeevi
- Yale School of Medicine, New Haven, CT
| | | | | | | | | | | | | | | |
Collapse
|
9
|
Patel AU, Shaker N, Mohanty S, Sharma S, Gangal S, Eloy C, Parwani AV. Cultivating Clinical Clarity through Computer Vision: A Current Perspective on Whole Slide Imaging and Artificial Intelligence. Diagnostics (Basel) 2022; 12:diagnostics12081778. [PMID: 35892487 PMCID: PMC9332710 DOI: 10.3390/diagnostics12081778] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/19/2022] [Revised: 07/10/2022] [Accepted: 07/11/2022] [Indexed: 11/17/2022] Open
Abstract
Diagnostic devices, methodological approaches, and traditional constructs of clinical pathology practice, cultivated throughout centuries, have transformed radically in the wake of explosive technological growth and other, e.g., environmental, catalysts of change. Ushered into the fray of modern laboratory medicine are digital imaging devices and machine-learning (ML) software fashioned to mitigate challenges, e.g., practitioner shortage while preparing clinicians for emerging interconnectivity of environments and diagnostic information in the era of big data. As computer vision shapes new constructs for the modern world and intertwines with clinical medicine, cultivating clarity of our new terrain through examining the trajectory and current scope of computational pathology and its pertinence to clinical practice is vital. Through review of numerous studies, we find developmental efforts for ML migrating from research to standardized clinical frameworks while overcoming obstacles that have formerly curtailed adoption of these tools, e.g., generalizability, data availability, and user-friendly accessibility. Groundbreaking validatory efforts have facilitated the clinical deployment of ML tools demonstrating the capacity to effectively aid in distinguishing tumor subtype and grade, classify early vs. advanced cancer stages, and assist in quality control and primary diagnosis applications. Case studies have demonstrated the benefits of streamlined, digitized workflows for practitioners alleviated by decreased burdens.
Collapse
Affiliation(s)
- Ankush U. Patel
- Mayo Clinic Department of Laboratory Medicine and Pathology, Rochester, MN 55905, USA
- Correspondence: ; Tel.: +1-206-451-3519
| | - Nada Shaker
- Department of Pathology, Wexner Medical Center, The Ohio State University, Columbus, OH 43210, USA; (N.S.); (S.G.); (A.V.P.)
| | - Sambit Mohanty
- CORE Diagnostics, Gurugram 122016, India; (S.M.); (S.S.)
- Advanced Medical Research Institute, Bareilly 243001, India
| | - Shivani Sharma
- CORE Diagnostics, Gurugram 122016, India; (S.M.); (S.S.)
| | - Shivam Gangal
- Department of Pathology, Wexner Medical Center, The Ohio State University, Columbus, OH 43210, USA; (N.S.); (S.G.); (A.V.P.)
- College of Engineering, Biomedical Engineering, The Ohio State University, Columbus, OH 43210, USA
| | - Catarina Eloy
- Institute of Molecular Pathology and Immunology of the University of Porto (IPATIMUP), Rua Júlio Amaral de Carvalho, 45, 4200-135 Porto, Portugal;
- Institute for Research and Innovation in Health (I3S Consortium), Rua Alfredo Allen, 208, 4200-135 Porto, Portugal
| | - Anil V. Parwani
- Department of Pathology, Wexner Medical Center, The Ohio State University, Columbus, OH 43210, USA; (N.S.); (S.G.); (A.V.P.)
- Cooperative Human Tissue Network (CHTN) Midwestern Division, Columbus, OH 43240, USA
| |
Collapse
|
10
|
Silva-Rodríguez J, Schmidt A, Sales MA, Molina R, Naranjo V. Proportion constrained weakly supervised histopathology image classification. Comput Biol Med 2022; 147:105714. [PMID: 35753089 DOI: 10.1016/j.compbiomed.2022.105714] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2022] [Revised: 05/14/2022] [Accepted: 06/04/2022] [Indexed: 11/16/2022]
Abstract
Multiple instance learning (MIL) deals with data grouped into bags of instances, of which only the global information is known. In recent years, this weakly supervised learning paradigm has become very popular in histological image analysis because it alleviates the burden of labeling all cancerous regions of large Whole Slide Images (WSIs) in detail. However, these methods require large datasets to perform properly, and many approaches only focus on simple binary classification. This often does not match the real-world problems where multi-label settings are frequent and possible constraints must be taken into account. In this work, we propose a novel multi-label MIL formulation based on inequality constraints that is able to incorporate prior knowledge about instance proportions. Our method has a theoretical foundation in optimization with log-barrier extensions, applied to bag-level class proportions. This encourages the model to respect the proportion ordering during training. Extensive experiments on a new public dataset of prostate cancer WSIs analysis, SICAP-MIL, demonstrate that using the prior proportion information we can achieve instance-level results similar to supervised methods on datasets of similar size. In comparison with prior MIL settings, our method allows for ∼13% improvements in instance-level accuracy, and ∼3% in the multi-label mean area under the ROC curve at the bag-level.
Collapse
Affiliation(s)
- Julio Silva-Rodríguez
- Institute of Transport and Territory, Universitat Politècnica de València, Valencia, Spain.
| | - Arne Schmidt
- Department of Computer Science and Artificial Intelligence, University of Granada, Granada, Spain.
| | - María A Sales
- Anatomical Pathology Service, University Clinical Hospital of Valencia, Valencia, Spain.
| | - Rafael Molina
- Department of Computer Science and Artificial Intelligence, University of Granada, Granada, Spain.
| | - Valery Naranjo
- Institute of Research and Innovation in Bioengineering, Universitat Politècnica de València, Valencia, Spain.
| |
Collapse
|
11
|
López-Pérez M, Schmidt A, Wu Y, Molina R, Katsaggelos AK. Deep Gaussian processes for multiple instance learning: Application to CT intracranial hemorrhage detection. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 219:106783. [PMID: 35390723 DOI: 10.1016/j.cmpb.2022.106783] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/29/2021] [Revised: 03/11/2022] [Accepted: 03/28/2022] [Indexed: 06/14/2023]
Abstract
BACKGROUND AND OBJECTIVE Intracranial hemorrhage (ICH) is a life-threatening emergency that can lead to brain damage or death, with high rates of mortality and morbidity. The fast and accurate detection of ICH is important for the patient to get an early and efficient treatment. To improve this diagnostic process, the application of Deep Learning (DL) models on head CT scans is an active area of research. Although promising results have been obtained, many of the proposed models require slice-level annotations by radiologists, which are costly and time-consuming. METHODS We formulate the ICH detection as a problem of Multiple Instance Learning (MIL) that allows training with only scan-level annotations. We develop a new probabilistic method based on Deep Gaussian Processes (DGP) that is able to train with this MIL setting and accurately predict ICH at both slice- and scan-level. The proposed DGPMIL model is able to capture complex feature relations by using multiple Gaussian Process (GP) layers, as we show experimentally. RESULTS To highlight the advantages of DGPMIL in a general MIL setting, we first conduct several controlled experiments on the MNIST dataset. We show that multiple GP layers outperform one-layer GP models, especially for complex feature distributions. For ICH detection experiments, we use two public brain CT datasets (RSNA and CQ500). We first train a Convolutional Neural Network (CNN) with an attention mechanism to extract the image features, which are fed into our DGPMIL model to perform the final predictions. The results show that DGPMIL model outperforms VGPMIL as well as the attention-based CNN for MIL and other state-of-the-art methods for this problem. The best performing DGPMIL model reaches an AUC-ROC of 0.957 (resp. 0.909) and an AUC-PR of 0.961 (resp. 0.889) on the RSNA (resp. CQ500) dataset. CONCLUSION The competitive performance at slice- and scan-level shows that DGPMIL model provides an accurate diagnosis on slices without the need for slice-level annotations by radiologists during training. As MIL is a common problem setting, our model can be applied to a broader range of other tasks, especially in medical image classification, where it can help the diagnostic process.
Collapse
Affiliation(s)
- Miguel López-Pérez
- Department of Computer Science and Artificial Intelligence, University of Granada, Granada 18010, Spain.
| | - Arne Schmidt
- Department of Computer Science and Artificial Intelligence, University of Granada, Granada 18010, Spain.
| | - Yunan Wu
- Department of Electrical Computer Engineering, Northwestern University, Evanston, IL, 60208 USA.
| | - Rafael Molina
- Department of Computer Science and Artificial Intelligence, University of Granada, Granada 18010, Spain.
| | - Aggelos K Katsaggelos
- Department of Electrical Computer Engineering, Northwestern University, Evanston, IL, 60208 USA.
| |
Collapse
|
12
|
Pérez-Bueno F, Serra JG, Vega M, Mateos J, Molina R, Katsaggelos AK. Bayesian K-SVD for H and E blind color deconvolution. Applications to stain normalization, data augmentation and cancer classification. Comput Med Imaging Graph 2022; 97:102048. [DOI: 10.1016/j.compmedimag.2022.102048] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/12/2021] [Revised: 11/04/2021] [Accepted: 02/05/2022] [Indexed: 12/17/2022]
|
13
|
Pérez-Bueno F, Vega M, Sales MA, Aneiros-Fernández J, Naranjo V, Molina R, Katsaggelos AK. Blind color deconvolution, normalization, and classification of histological images using general super Gaussian priors and Bayesian inference. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2021; 211:106453. [PMID: 34649072 DOI: 10.1016/j.cmpb.2021.106453] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/06/2021] [Accepted: 10/01/2021] [Indexed: 06/13/2023]
Abstract
BACKGROUND AND OBJECTIVE Color variations in digital histopathology severely impact the performance of computer-aided diagnosis systems. They are due to differences in the staining process and acquisition system, among other reasons. Blind color deconvolution techniques separate multi-stained images into single stained bands which, once normalized, can be used to eliminate these negative color variations and improve the performance of machine learning tasks. METHODS In this work, we decompose the observed RGB image in its hematoxylin and eosin components. We apply Bayesian modeling and inference based on the use of Super Gaussian sparse priors for each stain together with prior closeness to a given reference color-vector matrix. The hematoxylin and eosin components are then used for image normalization and classification of histological images. The proposed framework is tested on stain separation, image normalization, and cancer classification problems. The results are measured using the peak signal to noise ratio, normalized median intensity and the area under ROC curve on five different databases. RESULTS The obtained results show the superiority of our approach to current state-of-the-art blind color deconvolution techniques. In particular, the fidelity to the tissue improves 1,27 dB in mean PSNR. The normalized median intensity shows a good normalization quality of the proposed approach on the tested datasets. Finally, in cancer classification experiments the area under the ROC curve improves from 0.9491 to 0.9656 and from 0.9279 to 0.9541 on Camelyon-16 and Camelyon-17, respectively, when the original and processed images are used. Furthermore, these figures of merits are better than those obtained by the methods compared with. CONCLUSIONS The proposed framework for blind color deconvolution, normalization and classification of images guarantees fidelity to the tissue structure and can be used both for normalization and classification. In addition, color deconvolution enables the use of the optical density space for classification, which improves the classification performance.
Collapse
Affiliation(s)
- Fernando Pérez-Bueno
- Dpto. Ciencias de la Computación e Inteligencia Artificial, Universidad de Granada, Spain.
| | - Miguel Vega
- Dpto. de Lenguajes y Sistemas Informáticos, Universidad de Granada, Spain.
| | - María A Sales
- Anatomical Pathology Service, University Clinical Hospital of Valencia, Valencia, Spain.
| | - José Aneiros-Fernández
- Intercenter Unit of Pathological Anatomy, San Cecilio University Hospital, Granada, Spain.
| | - Valery Naranjo
- Dpto. de Comunicaciones, Universidad Politécnica de Valencia, Spain.
| | - Rafael Molina
- Dpto. Ciencias de la Computación e Inteligencia Artificial, Universidad de Granada, Spain.
| | - Aggelos K Katsaggelos
- Dept. of Electrical Engineering and Computer Science, Northwestern University, Evanston, IL, USA.
| |
Collapse
|
14
|
Silva-Rodriguez J, Colomer A, Dolz J, Naranjo V. Self-Learning for Weakly Supervised Gleason Grading of Local Patterns. IEEE J Biomed Health Inform 2021; 25:3094-3104. [PMID: 33621184 DOI: 10.1109/jbhi.2021.3061457] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
Prostate cancer is one of the main diseases affecting men worldwide. The gold standard for diagnosis and prognosis is the Gleason grading system. In this process, pathologists manually analyze prostate histology slides under microscope, in a high time-consuming and subjective task. In the last years, computer-aided-diagnosis (CAD) systems have emerged as a promising tool that could support pathologists in the daily clinical practice. Nevertheless, these systems are usually trained using tedious and prone-to-error pixel-level annotations of Gleason grades in the tissue. To alleviate the need of manual pixel-wise labeling, just a handful of works have been presented in the literature. Furthermore, despite the promising results achieved on global scoring the location of cancerous patterns in the tissue is only qualitatively addressed. These heatmaps of tumor regions, however, are crucial to the reliability of CAD systems as they provide explainability to the system's output and give confidence to pathologists that the model is focusing on medical relevant features. Motivated by this, we propose a novel weakly-supervised deep-learning model, based on self-learning CNNs, that leverages only the global Gleason score of gigapixel whole slide images during training to accurately perform both, grading of patch-level patterns and biopsy-level scoring. To evaluate the performance of the proposed method, we perform extensive experiments on three different external datasets for the patch-level Gleason grading, and on two different test sets for global Grade Group prediction. We empirically demonstrate that our approach outperforms its supervised counterpart on patch-level Gleason grading by a large margin, as well as state-of-the-art methods on global biopsy-level scoring. Particularly, the proposed model brings an average improvement on the Cohen's quadratic kappa ( κ) score of nearly 18% compared to full-supervision for the patch-level Gleason grading task. This suggests that the absence of the annotator's bias in our approach and the capability of using large weakly labeled datasets during training leads to higher performing and more robust models. Furthermore, raw features obtained from the patch-level classifier showed to generalize better than previous approaches in the literature to the subjective global biopsy-level scoring.
Collapse
|
15
|
DDV: A Taxonomy for Deep Learning Methods in Detecting Prostate Cancer. Neural Process Lett 2021. [DOI: 10.1007/s11063-021-10485-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
16
|
Oszwald A, Wasinger G, Pradere B, Shariat SF, Compérat EM. Artificial intelligence in prostate histopathology: where are we in 2021? Curr Opin Urol 2021; 31:430-435. [PMID: 33965977 DOI: 10.1097/mou.0000000000000883] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/20/2023]
Abstract
PURPOSE OF REVIEW Artificial intelligence has made an entrance into mainstream applications of daily life but the clinical deployment of artificial intelligence-supported histological analysis is still at infancy. Recent years have seen a surge in technological advance regarding the use of artificial intelligence in pathology, in particular in the diagnosis of prostate cancer. RECENT FINDINGS We review first impressions of how artificial intelligence impacts the clinical performance of pathologists in the analysis of prostate tissue. Several challenges in the deployment of artificial intelligence remain to be overcome. Finally, we discuss how artificial intelligence can help in generating new knowledge that is interpretable by humans. SUMMARY It is evident that artificial intelligence has the potential to outperform most pathologists in detecting prostate cancer, and does not suffer from inherent interobserver variability. Nonetheless, large clinical validation studies that unequivocally prove the benefit of artificial intelligence support in pathology are necessary. Regardless, artificial intelligence may soon automate and standardize many facets of routine work, including qualitative (i.e. Gleason Grading) and quantitative measures (i.e. portion of Gleason Grades and tumor volume). For the near future, a model where pathologists are enhanced by second-review or real-time artificial intelligence systems appears to be the most promising approach.
Collapse
Affiliation(s)
| | | | - Benjamin Pradere
- Department of Urology, Medical University of Vienna, Vienna, Austria
| | | | | |
Collapse
|
17
|
Epstein JI, Amin MB, Fine SW, Algaba F, Aron M, Baydar DE, Beltran AL, Brimo F, Cheville JC, Colecchia M, Comperat E, da Cunha IW, Delprado W, DeMarzo AM, Giannico GA, Gordetsky JB, Guo CC, Hansel DE, Hirsch MS, Huang J, Humphrey PA, Jimenez RE, Khani F, Kong Q, Kryvenko ON, Kunju LP, Lal P, Latour M, Lotan T, Maclean F, Magi-Galluzzi C, Mehra R, Menon S, Miyamoto H, Montironi R, Netto GJ, Nguyen JK, Osunkoya AO, Parwani A, Robinson BD, Rubin MA, Shah RB, So JS, Takahashi H, Tavora F, Tretiakova MS, True L, Wobker SE, Yang XJ, Zhou M, Zynger DL, Trpkov K. The 2019 Genitourinary Pathology Society (GUPS) White Paper on Contemporary Grading of Prostate Cancer. Arch Pathol Lab Med 2021; 145:461-493. [PMID: 32589068 DOI: 10.5858/arpa.2020-0015-ra] [Citation(s) in RCA: 136] [Impact Index Per Article: 45.3] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/28/2020] [Indexed: 11/06/2022]
Abstract
CONTEXT.— Controversies and uncertainty persist in prostate cancer grading. OBJECTIVE.— To update grading recommendations. DATA SOURCES.— Critical review of the literature along with pathology and clinician surveys. CONCLUSIONS.— Percent Gleason pattern 4 (%GP4) is as follows: (1) report %GP4 in needle biopsy with Grade Groups (GrGp) 2 and 3, and in needle biopsy on other parts (jars) of lower grade in cases with at least 1 part showing Gleason score (GS) 4 + 4 = 8; and (2) report %GP4: less than 5% or less than 10% and 10% increments thereafter. Tertiary grade patterns are as follows: (1) replace "tertiary grade pattern" in radical prostatectomy (RP) with "minor tertiary pattern 5 (TP5)," and only use in RP with GrGp 2 or 3 with less than 5% Gleason pattern 5; and (2) minor TP5 is noted along with the GS, with the GrGp based on the GS. Global score and magnetic resonance imaging (MRI)-targeted biopsies are as follows: (1) when multiple undesignated cores are taken from a single MRI-targeted lesion, an overall grade for that lesion is given as if all the involved cores were one long core; and (2) if providing a global score, when different scores are found in the standard and the MRI-targeted biopsy, give a single global score (factoring both the systematic standard and the MRI-targeted positive cores). Grade Groups are as follows: (1) Grade Groups (GrGp) is the terminology adopted by major world organizations; and (2) retain GS 3 + 5 = 8 in GrGp 4. Cribriform carcinoma is as follows: (1) report the presence or absence of cribriform glands in biopsy and RP with Gleason pattern 4 carcinoma. Intraductal carcinoma (IDC-P) is as follows: (1) report IDC-P in biopsy and RP; (2) use criteria based on dense cribriform glands (>50% of the gland is composed of epithelium relative to luminal spaces) and/or solid nests and/or marked pleomorphism/necrosis; (3) it is not necessary to perform basal cell immunostains on biopsy and RP to identify IDC-P if the results would not change the overall (highest) GS/GrGp part per case; (4) do not include IDC-P in determining the final GS/GrGp on biopsy and/or RP; and (5) "atypical intraductal proliferation (AIP)" is preferred for an intraductal proliferation of prostatic secretory cells which shows a greater degree of architectural complexity and/or cytological atypia than typical high-grade prostatic intraepithelial neoplasia, yet falling short of the strict diagnostic threshold for IDC-P. Molecular testing is as follows: (1) Ki67 is not ready for routine clinical use; (2) additional studies of active surveillance cohorts are needed to establish the utility of PTEN in this setting; and (3) dedicated studies of RNA-based assays in active surveillance populations are needed to substantiate the utility of these expensive tests in this setting. Artificial intelligence and novel grading schema are as follows: (1) incorporating reactive stromal grade, percent GP4, minor tertiary GP5, and cribriform/intraductal carcinoma are not ready for adoption in current practice.
Collapse
Affiliation(s)
- Jonathan I Epstein
- From the Departments of Pathology (Epstein, DeMarzo, Lotan), McGill University Health Center, Montréal, Quebec, Canada.,Urology (Epstein), David Geffen School of Medicine at UCLA, Los Angeles, California (Huang).,and Oncology (Epstein), The Johns Hopkins Medical Institutions, Baltimore, Maryland
| | - Mahul B Amin
- Department of Pathology and Laboratory Medicine and Urology, University of Tennessee Health Science, Memphis (Amin)
| | - Samson W Fine
- Department of Pathology, Memorial Sloan Kettering Cancer Center, New York, New York (Fine)
| | - Ferran Algaba
- Department of Pathology, Fundacio Puigvert, Barcelona, Spain (Algaba)
| | - Manju Aron
- Department of Pathology, University of Southern California, Los Angeles (Aron)
| | - Dilek E Baydar
- Department of Pathology, Faculty of Medicine, Koç University, İstanbul, Turkey (Baydar)
| | - Antonio Lopez Beltran
- Department of Pathology, Champalimaud Centre for the Unknown, Lisbon, Portugal (Beltran)
| | - Fadi Brimo
- Department of Pathology, McGill University Health Center, Montréal, Quebec, Canada (Brimo)
| | - John C Cheville
- Department of Pathology, Mayo Clinic, Rochester, Minnesota (Cheville, Jimenez)
| | - Maurizio Colecchia
- Department of Pathology and Laboratory Medicine, Fondazione IRCCS Istituto Nazionale dei Tumori, Milan, Italy (Colecchia)
| | - Eva Comperat
- Department of Pathology, Hôpital Tenon, Sorbonne University, Paris, France (Comperat)
| | | | | | - Angelo M DeMarzo
- From the Departments of Pathology (Epstein, DeMarzo, Lotan), McGill University Health Center, Montréal, Quebec, Canada
| | - Giovanna A Giannico
- Department of Pathology, Microbiology, and Immunology, Vanderbilt University Medical Center, Nashville, Tennessee (Giannico, Gordetsky)
| | - Jennifer B Gordetsky
- Department of Pathology, Microbiology, and Immunology, Vanderbilt University Medical Center, Nashville, Tennessee (Giannico, Gordetsky)
| | - Charles C Guo
- Department of Pathology, The University of Texas MD Anderson Cancer Center, Houston (Guo)
| | - Donna E Hansel
- Department of Pathology, Oregon Health and Science University, Portland (Hansel)
| | - Michelle S Hirsch
- Department of Pathology, Brigham and Women's Hospital, Harvard Medical School, Boston, Massachusetts (Hirsch)
| | - Jiaoti Huang
- Department of Pathology and Laboratory Medicine, David Geffen School of Medicine at UCLA, Los Angeles, California (Huang)
| | - Peter A Humphrey
- Department of Pathology, Yale School of Medicine, New Haven, Connecticut (Humphrey)
| | - Rafael E Jimenez
- Department of Pathology, Mayo Clinic, Rochester, Minnesota (Cheville, Jimenez)
| | - Francesca Khani
- Department of Pathology and Laboratory Medicine and Urology, Weill Cornell Medicine, New York, New York (Khani, Robinson)
| | - Qingnuan Kong
- Department of Pathology, Qingdao Municipal Hospital, Qingdao, Shandong, China (Kong).,Kong is currently located at Kaiser Permanente Sacramento Medical Center, Sacramento, California
| | - Oleksandr N Kryvenko
- Departments of Pathology and Laboratory Medicine and Sylvester Comprehensive Cancer Center, University of Miami Miller School of Medicine, Miami, Florida (Kryvenko)
| | - L Priya Kunju
- Department of Pathology, University of Michigan Medical School, Ann Arbor, Michigan (Kunju, Mehra)
| | - Priti Lal
- Perelman School of Medicine, University of Pennsylvania, Philadelphia (Lal)
| | - Mathieu Latour
- Department of Pathology, CHUM, Université de Montréal, Montréal, Quebec, Canada (Latour)
| | - Tamara Lotan
- From the Departments of Pathology (Epstein, DeMarzo, Lotan), McGill University Health Center, Montréal, Quebec, Canada
| | - Fiona Maclean
- Douglass Hanly Moir Pathology, Faculty of Medicine and Health Sciences Macquarie University, North Ryde, Australia (Maclean)
| | - Cristina Magi-Galluzzi
- Department of Pathology, The University of Alabama at Birmingham, Birmingham (Magi-Galluzzi, Netto)
| | - Rohit Mehra
- Department of Pathology, University of Michigan Medical School, Ann Arbor, Michigan (Kunju, Mehra)
| | - Santosh Menon
- Department of Surgical Pathology, Tata Memorial Hospital, Parel, Mumbai, India (Menon)
| | - Hiroshi Miyamoto
- Departments of Pathology and Laboratory Medicine and Urology, University of Rochester Medical Center, Rochester, New York (Miyamoto)
| | - Rodolfo Montironi
- Section of Pathological Anatomy, School of Medicine, Polytechnic University of the Marche Region, United Hospitals, Ancona, Italy (Montironi)
| | - George J Netto
- Department of Pathology, The University of Alabama at Birmingham, Birmingham (Magi-Galluzzi, Netto)
| | - Jane K Nguyen
- Robert J. Tomsich Pathology and Laboratory Medicine Institute, Cleveland Clinic, Cleveland, Ohio (Nguyen)
| | - Adeboye O Osunkoya
- Department of Pathology, Emory University School of Medicine, Atlanta, Georgia (Osunkoya)
| | - Anil Parwani
- Department of Pathology, Ohio State University, Columbus (Parwani, Zynger)
| | - Brian D Robinson
- Department of Pathology and Laboratory Medicine and Urology, Weill Cornell Medicine, New York, New York (Khani, Robinson)
| | - Mark A Rubin
- Department for BioMedical Research, University of Bern, Bern, Switzerland (Rubin)
| | - Rajal B Shah
- Department of Pathology, The University of Texas Southwestern Medical Center, Dallas (Shah)
| | - Jeffrey S So
- Institute of Pathology, St Luke's Medical Center, Quezon City and Global City, Philippines (So)
| | - Hiroyuki Takahashi
- Department of Pathology, The Jikei University School of Medicine, Tokyo, Japan (Takahashi)
| | - Fabio Tavora
- Argos Laboratory, Federal University of Ceara, Fortaleza, Brazil (Tavora)
| | - Maria S Tretiakova
- Department of Pathology, University of Washington School of Medicine, Seattle (Tretiakova, True)
| | - Lawrence True
- Department of Pathology, University of Washington School of Medicine, Seattle (Tretiakova, True)
| | - Sara E Wobker
- Departments of Pathology and Laboratory Medicine and Urology, University of North Carolina, Chapel Hill (Wobker)
| | - Ximing J Yang
- Department of Pathology, Northwestern University, Chicago, Illinois (Yang)
| | - Ming Zhou
- Department of Pathology, Tufts Medical Center, Boston, Massachusetts (Zhou)
| | - Debra L Zynger
- Department of Pathology, Ohio State University, Columbus (Parwani, Zynger)
| | - Kiril Trpkov
- and Department of Pathology and Laboratory Medicine, University of Calgary, Calgary, Alberta, Canada (Trpkov)
| |
Collapse
|
18
|
Li J, Li W, Sisk A, Ye H, Wallace WD, Speier W, Arnold CW. A multi-resolution model for histopathology image classification and localization with multiple instance learning. Comput Biol Med 2021; 131:104253. [PMID: 33601084 DOI: 10.1016/j.compbiomed.2021.104253] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2020] [Revised: 01/31/2021] [Accepted: 02/03/2021] [Indexed: 12/17/2022]
Abstract
Large numbers of histopathological images have been digitized into high resolution whole slide images, opening opportunities in developing computational image analysis tools to reduce pathologists' workload and potentially improve inter- and intra-observer agreement. Most previous work on whole slide image analysis has focused on classification or segmentation of small pre-selected regions-of-interest, which requires fine-grained annotation and is non-trivial to extend for large-scale whole slide analysis. In this paper, we proposed a multi-resolution multiple instance learning model that leverages saliency maps to detect suspicious regions for fine-grained grade prediction. Instead of relying on expensive region- or pixel-level annotations, our model can be trained end-to-end with only slide-level labels. The model is developed on a large-scale prostate biopsy dataset containing 20,229 slides from 830 patients. The model achieved 92.7% accuracy, 81.8% Cohen's Kappa for benign, low grade (i.e. Grade group 1) and high grade (i.e. Grade group ≥ 2) prediction, an area under the receiver operating characteristic curve (AUROC) of 98.2% and an average precision (AP) of 97.4% for differentiating malignant and benign slides. The model obtained an AUROC of 99.4% and an AP of 99.8% for cancer detection on an external dataset.
Collapse
Affiliation(s)
- Jiayun Li
- Computational Diagnostics Lab, UCLA, 924 Westwood Blvd Suite 600, Los Angeles, CA, 90024, USA; Department of Radiology, UCLA, 924 Westwood Blvd Suite 600, Los Angeles, CA, 90024, USA.
| | - Wenyuan Li
- Computational Diagnostics Lab, UCLA, 924 Westwood Blvd Suite 600, Los Angeles, CA, 90024, USA; Department of Radiology, UCLA, 924 Westwood Blvd Suite 600, Los Angeles, CA, 90024, USA
| | - Anthony Sisk
- Department of Pathology & Laboratory Medicine, UCLA, 10833 Le Conte Ave, Los Angeles, CA, 90095, USA
| | - Huihui Ye
- Department of Pathology & Laboratory Medicine, UCLA, 10833 Le Conte Ave, Los Angeles, CA, 90095, USA
| | - W Dean Wallace
- Department of Pathology, USC, 2011 Zonal Avenue, Los Angeles, CA, 90033, USA
| | - William Speier
- Computational Diagnostics Lab, UCLA, 924 Westwood Blvd Suite 600, Los Angeles, CA, 90024, USA
| | - Corey W Arnold
- Computational Diagnostics Lab, UCLA, 924 Westwood Blvd Suite 600, Los Angeles, CA, 90024, USA; Department of Radiology, UCLA, 924 Westwood Blvd Suite 600, Los Angeles, CA, 90024, USA; Department of Pathology & Laboratory Medicine, UCLA, 10833 Le Conte Ave, Los Angeles, CA, 90095, USA.
| |
Collapse
|
19
|
Silva-Rodríguez J, Colomer A, Sales MA, Molina R, Naranjo V. Going deeper through the Gleason scoring scale: An automatic end-to-end system for histology prostate grading and cribriform pattern detection. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2020; 195:105637. [PMID: 32653747 DOI: 10.1016/j.cmpb.2020.105637] [Citation(s) in RCA: 41] [Impact Index Per Article: 10.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/09/2020] [Accepted: 06/28/2020] [Indexed: 06/11/2023]
Abstract
BACKGROUND AND OBJECTIVE Prostate cancer is one of the most common diseases affecting men worldwide. The Gleason scoring system is the primary diagnostic and prognostic tool for prostate cancer. Furthermore, recent reports indicate that the presence of patterns of the Gleason scale such as the cribriform pattern may also correlate with a worse prognosis compared to other patterns belonging to the Gleason grade 4. Current clinical guidelines have indicated the convenience of highlight its presence during the analysis of biopsies. All these requirements suppose a great workload for the pathologist during the analysis of each sample, which is based on the pathologist's visual analysis of the morphology and organisation of the glands in the tissue, a time-consuming and subjective task. In recent years, with the development of digitisation devices, the use of computer vision techniques for the analysis of biopsies has increased. However, to the best of the authors' knowledge, the development of algorithms to automatically detect individual cribriform patterns belonging to Gleason grade 4 has not yet been studied in the literature. The objective of the work presented in this paper is to develop a deep-learning-based system able to support pathologists in the daily analysis of prostate biopsies. This analysis must include the Gleason grading of local structures, the detection of cribriform patterns, and the Gleason scoring of the whole biopsy. METHODS The methodological core of this work is a patch-wise predictive model based on convolutional neural networks able to determine the presence of cancerous patterns based on the Gleason grading system. In particular, we train from scratch a simple self-design architecture with three filters and a top model with global-max pooling. The cribriform pattern is detected by retraining the set of filters of the last convolutional layer in the network. Subsequently, a biopsy-level prediction map is reconstructed by bi-linear interpolation of the patch-level prediction of the Gleason grades. In addition, from the reconstructed prediction map, we compute the percentage of each Gleason grade in the tissue to feed a multi-layer perceptron which provides a biopsy-level score. RESULTS In our SICAPv2 database, composed of 182 annotated whole slide images, we obtained a Cohen's quadratic kappa of 0.77 in the test set for the patch-level Gleason grading with the proposed architecture trained from scratch. Our results outperform previous ones reported in the literature. Furthermore, this model reaches the level of fine-tuned state-of-the-art architectures in a patient-based four groups cross validation. In the cribriform pattern detection task, we obtained an area under ROC curve of 0.82. Regarding the biopsy Gleason scoring, we achieved a quadratic Cohen's Kappa of 0.81 in the test subset. Shallow CNN architectures trained from scratch outperform current state-of-the-art methods for Gleason grades classification. Our proposed model is capable of characterising the different Gleason grades in prostate tissue by extracting low-level features through three basic blocks (i.e. convolutional layer + max pooling). The use of global-max pooling to reduce each activation map has shown to be a key factor for reducing complexity in the model and avoiding overfitting. Regarding the Gleason scoring of biopsies, a multi-layer perceptron has shown to better model the decision-making of pathologists than previous simpler models used in the literature.
Collapse
Affiliation(s)
- Julio Silva-Rodríguez
- Institute of Transport and Territory, Universitat Politècnica de València, Valencia, Spain.
| | - Adrián Colomer
- Institute of Research and Innovation in Bioengineering, Universitat Politècnica de València, Valencia, Spain.
| | - María A Sales
- Anatomical Pathology Service, University Clinical Hospital of Valencia, Valencia, Spain.
| | - Rafael Molina
- Department of Computer Science and Artificial Intelligence, University of Granada, Granada, Spain.
| | - Valery Naranjo
- Institute of Research and Innovation in Bioengineering, Universitat Politècnica de València, Valencia, Spain.
| |
Collapse
|