1
|
Fayyad J, Alijani S, Najjaran H. Empirical validation of Conformal Prediction for trustworthy skin lesions classification. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 253:108231. [PMID: 38820714 DOI: 10.1016/j.cmpb.2024.108231] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/17/2024] [Revised: 03/15/2024] [Accepted: 05/14/2024] [Indexed: 06/02/2024]
Abstract
BACKGROUND AND OBJECTIVE Uncertainty quantification is a pivotal field that contributes to realizing reliable and robust systems. It becomes instrumental in fortifying safe decisions by providing complementary information, particularly within high-risk applications. existing studies have explored various methods that often operate under specific assumptions or necessitate substantial modifications to the network architecture to effectively account for uncertainties. The objective of this paper is to study Conformal Prediction, an emerging distribution-free uncertainty quantification technique, and provide a comprehensive understanding of the advantages and limitations inherent in various methods within the medical imaging field. METHODS In this study, we developed Conformal Prediction, Monte Carlo Dropout, and Evidential Deep Learning approaches to assess uncertainty quantification in deep neural networks. The effectiveness of these methods is evaluated using three public medical imaging datasets focused on detecting pigmented skin lesions and blood cell types. RESULTS The experimental results demonstrate a significant enhancement in uncertainty quantification with the utilization of the Conformal Prediction method, surpassing the performance of the other two methods. Furthermore, the results present insights into the effectiveness of each uncertainty method in handling Out-of-Distribution samples from domain-shifted datasets. Our code is available at: github.com/jfayyad/ConformalDx. CONCLUSIONS Our conclusion highlights a robust and consistent performance of conformal prediction across diverse testing conditions. This positions it as the preferred choice for decision-making in safety-critical applications.
Collapse
|
2
|
Victor D, Amina B, Diederik C, Francesco C, Raphaël F, Damien E. Towards CO 2 valorization in a multi remote renewable energy hub framework with uncertainty quantification. JOURNAL OF ENVIRONMENTAL MANAGEMENT 2024; 363:121262. [PMID: 38850908 DOI: 10.1016/j.jenvman.2024.121262] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/10/2024] [Revised: 04/25/2024] [Accepted: 05/26/2024] [Indexed: 06/10/2024]
Abstract
In this paper, we propose a multi-RREH (Remote Renewable Energy Hub) based optimization framework. This framework allows a valorization of CO2 using carbon capture technologies. This valorization is grounded on the idea that CO2 gathered from the atmosphere or post combustion can be combined with hydrogen to produce synthetic methane. The hydrogen is obtained from water electrolysis using renewable energy. Such renewable energy is generated in RREH, which are locations where RE is cheap and abundant (e.g., solar PV in the Sahara Desert, or wind in Greenland). We instantiate our framework on a case study focusing on Belgium and 2 RREH, and we conduct a techno-economic analysis under uncertainty. This analysis highlights, among others, the interest in capturing CO2 via Post Combustion Carbon Capture (PCCC) rather than only through Direct Air Capture (DAC) for methane synthesis in RREH. By doing so, a notable reduction of 10% is observed in the total cost of the system under our reference scenario. In addition, we use our framework to derive a carbon price threshold above which carbon capture technologies may start playing a pivotal role in the decarbonation process of our industries.
Collapse
|
3
|
Mamis K, Farazmand M. Modeling correlated uncertainties in stochastic compartmental models. Math Biosci 2024:109226. [PMID: 38838933 DOI: 10.1016/j.mbs.2024.109226] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/24/2023] [Revised: 05/15/2024] [Accepted: 05/30/2024] [Indexed: 06/07/2024]
Abstract
We consider compartmental models of communicable disease with uncertain contact rates. Stochastic fluctuations are often added to the contact rate to account for uncertainties. White noise, which is the typical choice for the fluctuations, leads to significant underestimation of the disease severity. Here, starting from reasonable assumptions on the social behavior of individuals, we model the contacts as a Markov process which takes into account the temporal correlations present in human social activities. Consequently, we show that the mean-reverting Ornstein-Uhlenbeck (OU) process is the correct model for the stochastic contact rate. We demonstrate the implication of our model on two examples: a Susceptibles-Infected-Susceptibles (SIS) model and a Susceptibles-Exposed-Infected-Removed (SEIR) model of the COVID-19 pandemic and compare the results to the available US data from the Johns Hopkins University database. In particular, we observe that both compartmental models with white noise uncertainties undergo transitions that lead to the systematic underestimation of the spread of the disease. In contrast, modeling the contact rate with the OU process significantly hinders such unrealistic noise-induced transitions. For the SIS model, we derive its stationary probability density analytically, for both white and correlated noise. This allows us to give a complete description of the model's asymptotic behavior as a function of its bifurcation parameters, i.e., the basic reproduction number, noise intensity, and correlation time. For the SEIR model, where the probability density is not available in closed form, we study the transitions using Monte Carlo simulations. Our modeling approach can be used to quantify uncertain parameters in a broad range of biological systems.
Collapse
|
4
|
Ninno F, Chiastra C, Colombo M, Dardik A, Strosberg D, Aboian E, Tsui J, Bartlett M, Balabani S, Díaz-Zuccarini V. Modelling lower-limb peripheral arterial disease using clinically available datasets: impact of inflow boundary conditions on hemodynamic indices for restenosis prediction. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 251:108214. [PMID: 38759252 DOI: 10.1016/j.cmpb.2024.108214] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/15/2024] [Revised: 04/22/2024] [Accepted: 05/01/2024] [Indexed: 05/19/2024]
Abstract
BACKGROUND AND OBJECTIVES The integration of hemodynamic markers as risk factors in restenosis prediction models for lower-limb peripheral arteries is hindered by fragmented clinical datasets. Computed tomography (CT) scans enable vessel geometry reconstruction and can be obtained at different times than the Doppler ultrasound (DUS) images, which provide information on blood flow velocity. Computational fluid dynamics (CFD) simulations allow the computation of near-wall hemodynamic indices, whose accuracy depends on the prescribed inlet boundary condition (BC), derived from the DUS images. This study aims to: (i) investigate the impact of different DUS-derived velocity waveforms on CFD results; (ii) test whether the same vessel areas, subjected to altered hemodynamics, can be detected independently of the applied inlet BC; (iii) suggest suitable DUS images to obtain reliable CFD results. METHODS CFD simulations were conducted on three patients treated with bypass surgery, using patient-specific DUS-derived inlet BCs recorded at either the same or different time points than the CT scan. The impact of the chosen inflow condition on bypass hemodynamics was assessed in terms of wall shear stress (WSS)-derived quantities. Patient-specific critical thresholds for the hemodynamic indices were applied to identify critical luminal areas and compare the results with a reference obtained with a DUS image acquired in close temporal proximity to the CT scan. RESULTS The main findings indicate that: (i) DUS-derived inlet velocity waveforms acquired at different time points than the CT scan led to statistically significantly different CFD results (p<0.001); (ii) the same luminal surface areas, exposed to low time-averaged WSS, could be identified independently of the applied inlet BCs; (iii) similar outcomes were observed for the other hemodynamic indices if the prescribed inlet velocity waveform had the same shape and comparable systolic acceleration time to the one recorded in close temporal proximity to the CT scan. CONCLUSIONS Despite a lack of standardised data collection for diseased lower-limb peripheral arteries, an accurate estimation of luminal areas subjected to altered near-wall hemodynamics is possible independently of the applied inlet BC. This holds if the applied inlet waveform shares some characteristics - derivable from the DUS report - as one matching the acquisition time of the CT scan.
Collapse
|
5
|
Schäfer F, Sturdy J, Hellevik LR. Age and sex-dependent sensitivity analysis of a common carotid artery model. Biomech Model Mechanobiol 2024; 23:825-843. [PMID: 38369558 PMCID: PMC11101589 DOI: 10.1007/s10237-023-01808-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Accepted: 12/22/2023] [Indexed: 02/20/2024]
Abstract
The common carotid artery (CCA) is an accessible and informative site for assessing cardiovascular function which makes it a prime candidate for clinically relevant computational modelling. The interpretation of supplemental information possible through modelling is encumbered by measurement uncertainty and population variability in model parameters. The distribution of model parameters likely depends on the specific sub-population of interest and delineation based on sex, age or health status may correspond to distinct ranges of typical parameter values. To assess this impact in a 1D-CCA-model, we delineated specific sub-populations based on age, sex and health status and carried out uncertainty quantification and sensitivity analysis for each sub-population. We performed a structured literature review to characterize sub-population-specific variabilities for eight model parameters without consideration of health status; variations for a healthy sub-populations were based on previously established references values. The variabilities of diameter and distensibility found in the literature review differed from those previously established in a healthy population. Model diameter change and pulse pressure were most sensitive to variations in distensibility, while pressure was most sensitive to resistance in the Windkessel model for all groups. Uncertainties were lower when variabilities were based on a healthy sub-population; however, the qualitative distribution of sensitivity indices was largely similar between the healthy and general population. Average sensitivity of the pressure waveform showed a moderate dependence on age with decreasing sensitivity to distal resistance and increasing sensitivity to distensibility and diameter. The female population was less sensitive to variations in diameter but more sensitive to distensibility coefficient than the male population. Overall, as hypothesized input variabilities differed between sub-populations and resulted in distinct uncertainties and sensitivities of the 1D-CCA-model outputs, particularly over age for the pressure waveform and between males and females for pulse pressure.
Collapse
|
6
|
Huang L, Ruan S, Xing Y, Feng M. A review of uncertainty quantification in medical image analysis: Probabilistic and non-probabilistic methods. Med Image Anal 2024; 97:103223. [PMID: 38861770 DOI: 10.1016/j.media.2024.103223] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2023] [Revised: 03/16/2024] [Accepted: 05/27/2024] [Indexed: 06/13/2024]
Abstract
The comprehensive integration of machine learning healthcare models within clinical practice remains suboptimal, notwithstanding the proliferation of high-performing solutions reported in the literature. A predominant factor hindering widespread adoption pertains to an insufficiency of evidence affirming the reliability of the aforementioned models. Recently, uncertainty quantification methods have been proposed as a potential solution to quantify the reliability of machine learning models and thus increase the interpretability and acceptability of the results. In this review, we offer a comprehensive overview of the prevailing methods proposed to quantify the uncertainty inherent in machine learning models developed for various medical image tasks. Contrary to earlier reviews that exclusively focused on probabilistic methods, this review also explores non-probabilistic approaches, thereby furnishing a more holistic survey of research pertaining to uncertainty quantification for machine learning models. Analysis of medical images with the summary and discussion on medical applications and the corresponding uncertainty evaluation protocols are presented, which focus on the specific challenges of uncertainty in medical image analysis. We also highlight some potential future research work at the end. Generally, this review aims to allow researchers from both clinical and technical backgrounds to gain a quick and yet in-depth understanding of the research in uncertainty quantification for medical image analysis machine learning models.
Collapse
|
7
|
McCullough JWS, Coveney PV. Uncertainty quantification of the lattice Boltzmann method focussing on studies of human-scale vascular blood flow. Sci Rep 2024; 14:11317. [PMID: 38760455 PMCID: PMC11101457 DOI: 10.1038/s41598-024-61708-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2023] [Accepted: 05/08/2024] [Indexed: 05/19/2024] Open
Abstract
Uncertainty quantification is becoming a key tool to ensure that numerical models can be sufficiently trusted to be used in domains such as medical device design. Demonstration of how input parameters impact the quantities of interest generated by any numerical model is essential to understanding the limits of its reliability. With the lattice Boltzmann method now a widely used approach for computational fluid dynamics, building greater understanding of its numerical uncertainty characteristics will support its further use in science and industry. In this study we apply an in-depth uncertainty quantification study of the lattice Boltzmann method in a canonical bifurcating geometry that is representative of the vascular junctions present in arterial and venous domains. These campaigns examine how quantities of interest-pressure and velocity along the central axes of the bifurcation-are influenced by the algorithmic parameters of the lattice Boltzmann method and the parameters controlling the values imposed at inlet velocity and outlet pressure boundary conditions. We also conduct a similar campaign on a set of personalised vessels to further illustrate the application of these techniques. Our work provides insights into how input parameters and boundary conditions impact the velocity and pressure distributions calculated in a simulation and can guide the choices of such values when applied to vascular studies of patient specific geometries. We observe that, from an algorithmic perspective, the number of time steps and the size of the grid spacing are the most influential parameters. When considering the influence of boundary conditions, we note that the magnitude of the inlet velocity and the mean pressure applied within sinusoidal pressure outlets have the greatest impact on output quantities of interest. We also observe that, when comparing the magnitude of variation imposed in the input parameters with that observed in the output quantities, this variability is particularly magnified when the input velocity is altered. This study also demonstrates how open-source toolkits for validation, verification and uncertainty quantification can be applied to numerical models deployed on high-performance computers without the need for modifying the simulation code itself. Such an ability is key to the more widespread adoption of the analysis of uncertainty in numerical models by significantly reducing the complexity of their execution and analysis.
Collapse
|
8
|
Chen K, Li G, Li H, Wang Y, Wang W, Liu Q, Wang H. Quantifying uncertainty: Air quality forecasting based on dynamic spatial-temporal denoising diffusion probabilistic model. ENVIRONMENTAL RESEARCH 2024; 249:118438. [PMID: 38350546 DOI: 10.1016/j.envres.2024.118438] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/14/2024] [Revised: 02/04/2024] [Accepted: 02/05/2024] [Indexed: 02/15/2024]
Abstract
Air pollution constitutes a substantial peril to human health, thereby catalyzing the evolution of an array of air quality prediction models. These models span from mechanistic and statistical strategies to machine learning methodologies. The burgeoning field of deep learning has given rise to a plethora of advanced models, which have demonstrated commendable performance. However, previous investigations have overlooked the salience of quantifying prediction uncertainties and potential future interconnections among air monitoring stations. Moreover, prior research typically utilized static predetermined spatial relationships, neglecting dynamic dependencies. To address these limitations, we propose a model named Dynamic Spatial-Temporal Denoising Diffusion Probabilistic Model (DST-DDPM) for air quality prediction. Our model is underpinned by the renowned denoising diffusion model, aiding us in discerning indeterminacy. In order to encapsulate dynamic patterns, we design a dynamic context encoder to generate dynamic adjacency matrices, whilst maintaining static spatial information. Furthermore, we incorporate a spatial-temporal denoising model to concurrently learn both spatial and temporal dependencies. Authenticating our model's performance using a real-world dataset collected in Beijing, the outcomes indicate that our model eclipses other baseline models in terms of both short-term and long-term predictions by 1.36% and 11.62% respectively. Finally, we conduct a case study to exhibit our model's capacity to quantify uncertainties.
Collapse
|
9
|
Ma Q, Liu Z, Zhang T, Zhao S, Gao X, Sun T, Dai Y. Multielement simultaneous quantitative analysis of trace elements in stainless steel via full spectrum laser-induced breakdown spectroscopy. Talanta 2024; 272:125745. [PMID: 38367401 DOI: 10.1016/j.talanta.2024.125745] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2023] [Revised: 01/16/2024] [Accepted: 02/05/2024] [Indexed: 02/19/2024]
Abstract
Laser-Induced Breakdown Spectroscopy (LIBS) instruments are increasingly recognized as valuable tools for detecting trace metal elements due to their simplicity, rapid detection, and ability to perform simultaneous multi-element analysis. Traditional LIBS modeling often relies on empirical or machine learning-based feature band selection to establish quantitative models. In this study, we introduce a novel approach-simultaneous multi-element quantitative analysis based on the entire spectrum, which enhances model establishment efficiency and leverages the advantages of LIBS. By logarithmically processing the spectra and quantifying the cognitive uncertainty of the model, we achieved remarkable predictive performance (R2) for trace elements Mn, Mo, Cr, and Cu (0.9876, 0.9879, 0.9891, and 0.9841, respectively) in stainless steel. Our multi-element model shares features and parameters during the learning process, effectively mitigating the impact of matrix effects and self-absorption. Additionally, we introduce a cognitive error term to quantify the cognitive uncertainty of the model. The results suggest that our approach has significant potential in the quantitative analysis of trace elements, providing a reliable data processing method for efficient and accurate multi-task analysis in LIBS. This methodology holds promising applications in the field of LIBS quantitative analysis.
Collapse
|
10
|
Li L, Chang J, Vakanski A, Wang Y, Yao T, Xian M. Uncertainty quantification in multivariable regression for material property prediction with Bayesian neural networks. Sci Rep 2024; 14:10543. [PMID: 38719870 PMCID: PMC11078957 DOI: 10.1038/s41598-024-61189-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2023] [Accepted: 05/02/2024] [Indexed: 05/12/2024] Open
Abstract
With the increased use of data-driven approaches and machine learning-based methods in material science, the importance of reliable uncertainty quantification (UQ) of the predicted variables for informed decision-making cannot be overstated. UQ in material property prediction poses unique challenges, including multi-scale and multi-physics nature of materials, intricate interactions between numerous factors, limited availability of large curated datasets, etc. In this work, we introduce a physics-informed Bayesian Neural Networks (BNNs) approach for UQ, which integrates knowledge from governing laws in materials to guide the models toward physically consistent predictions. To evaluate the approach, we present case studies for predicting the creep rupture life of steel alloys. Experimental validation with three datasets of creep tests demonstrates that this method produces point predictions and uncertainty estimations that are competitive or exceed the performance of conventional UQ methods such as Gaussian Process Regression. Additionally, we evaluate the suitability of employing UQ in an active learning scenario and report competitive performance. The most promising framework for creep life prediction is BNNs based on Markov Chain Monte Carlo approximation of the posterior distribution of network parameters, as it provided more reliable results in comparison to BNNs based on variational inference approximation or related NNs with probabilistic outputs.
Collapse
|
11
|
Cui T, De Sterck H, Gilbert AD, Polishchuk S, Scheichl R. Multilevel Monte Carlo Methods for Stochastic Convection-Diffusion Eigenvalue Problems. JOURNAL OF SCIENTIFIC COMPUTING 2024; 99:77. [PMID: 38708025 PMCID: PMC11068587 DOI: 10.1007/s10915-024-02539-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/04/2023] [Revised: 12/21/2023] [Accepted: 03/14/2024] [Indexed: 05/07/2024]
Abstract
We develop new multilevel Monte Carlo (MLMC) methods to estimate the expectation of the smallest eigenvalue of a stochastic convection-diffusion operator with random coefficients. The MLMC method is based on a sequence of finite element (FE) discretizations of the eigenvalue problem on a hierarchy of increasingly finer meshes. For the discretized, algebraic eigenproblems we use both the Rayleigh quotient (RQ) iteration and implicitly restarted Arnoldi (IRA), providing an analysis of the cost in each case. By studying the variance on each level and adapting classical FE error bounds to the stochastic setting, we are able to bound the total error of our MLMC estimator and provide a complexity analysis. As expected, the complexity bound for our MLMC estimator is superior to plain Monte Carlo. To improve the efficiency of the MLMC further, we exploit the hierarchy of meshes and use coarser approximations as starting values for the eigensolvers on finer ones. To improve the stability of the MLMC method for convection-dominated problems, we employ two additional strategies. First, we consider the streamline upwind Petrov-Galerkin formulation of the discrete eigenvalue problem, which allows us to start the MLMC method on coarser meshes than is possible with standard FEs. Second, we apply a homotopy method to add stability to the eigensolver for each sample. Finally, we present a multilevel quasi-Monte Carlo method that replaces Monte Carlo with a quasi-Monte Carlo (QMC) rule on each level. Due to the faster convergence of QMC, this improves the overall complexity. We provide detailed numerical results comparing our different strategies to demonstrate the practical feasibility of the MLMC method in different use cases. The results support our complexity analysis and further demonstrate the superiority over plain Monte Carlo in all cases.
Collapse
|
12
|
Nagaraja S, Loughran G, Baumann AP, Kartikeya K, Horner M. Establishing finite element model credibility of a pedicle screw system under compression-bending: An end-to-end example of the ASME V&V 40 standard. Methods 2024; 225:74-88. [PMID: 38493931 DOI: 10.1016/j.ymeth.2024.03.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2024] [Revised: 03/06/2024] [Accepted: 03/08/2024] [Indexed: 03/19/2024] Open
Abstract
Computational modeling and simulation (CM&S) is a key tool in medical device design, development, and regulatory approval. For example, finite element analysis (FEA) is widely used to understand the mechanical integrity and durability of orthopaedic implants. The ASME V&V 40 standard and supporting FDA guidance provide a framework for establishing model credibility, enabling deeper reliance on CM&S throughout the total product lifecycle. Examples of how to apply the principles outlined in the ASME V&V 40 standard are important to facilitating greater adoption by the medical device community, but few published examples are available that demonstrate best practices. Therefore, this paper outlines an end-to-end (E2E) example of the ASME V&V 40 standard applied to an orthopaedic implant. The objective of this study was to illustrate how to establish the credibility of a computational model intended for use as part of regulatory evaluation. In particular, this study focused on whether a design change to a spinal pedicle screw construct (specifically, the addition of a cannulation to an existing non-cannulated pedicle screw) would compromise the rod-screw construct mechanical performance. This question of interest (?OI) was addressed by establishing model credibility requirements according to the ASME V&V 40 standard. Experimental testing to support model validation was performed using spinal rods and non-cannulated pedicle screw constructs made with medical grade titanium (Ti-6Al-4V ELI). FEA replicating the experimental tests was performed by three independent modelers and validated through comparisons of common mechanical properties such as stiffness and yield force. The validated model was then used to simulate F1717 compression-bending testing on the new cannulated pedicle screw design to answer the ?OI, without performing any additional experimental testing. This E2E example provides a realistic scenario for the application of the ASME V&V 40 standard to orthopedic medical device applications.
Collapse
|
13
|
Blair J, Stephen B, Brown B, McArthur S, Gorman D, Forbes A, Pottier C, McAlorum J, Dow H, Perry M. Photometric stereo data for the validation of a structural health monitoring test rig. Data Brief 2024; 53:110164. [PMID: 38375140 PMCID: PMC10875225 DOI: 10.1016/j.dib.2024.110164] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Revised: 12/22/2023] [Accepted: 01/31/2024] [Indexed: 02/21/2024] Open
Abstract
Photometric stereo uses images of objects illuminated from various directions to calculate surface normals which can be used to generate 3D meshes of the object. Such meshes can be used by engineers to estimate damage of a concrete surface, or track damage progression over time to inform maintenance decisions. This dataset [1] was collected to quantify the uncertainty in a photometric stereo test rig through both the comparison with a well characterised method (coordinate measurement machine) and experiment virtualisation. Data was collected for 9 real objects using both the test rig and the coordinate measurement machine. These objects range from clay statues to damaged concrete slabs. Furthermore, synthetic data for 12 objects was created via virtual renders generated using Blender (3D software) [2]. The two methods of data generation allowed the decoupling of the physical rig (used to light and photograph objects) and the photometric stereo algorithm (used to convert images and lighting information into 3D meshes). This data can allow users to: test their own photometric stereo algorithms, with specialised data created for structural health monitoring applications; provide an industrially relevant case study to develop and test uncertainty quantification methods on test rigs for structural health monitoring of concrete; or develop data processing methodologies for the alignment of scaled, translated, and rotated data.
Collapse
|
14
|
Zahari R, Cox J, Obara B. Uncertainty-aware image classification on 3D CT lung. Comput Biol Med 2024; 172:108324. [PMID: 38508053 DOI: 10.1016/j.compbiomed.2024.108324] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2023] [Revised: 03/06/2024] [Accepted: 03/14/2024] [Indexed: 03/22/2024]
Abstract
Early detection is crucial for lung cancer to prolong the patient's survival. Existing model architectures used in such systems have shown promising results. However, they lack reliability and robustness in their predictions and the models are typically evaluated on a single dataset, making them overconfident when a new class is present. With the existence of uncertainty, uncertain images can be referred to medical experts for a second opinion. Thus, we propose an uncertainty-aware framework that includes three phases: data preprocessing and model selection and evaluation, uncertainty quantification (UQ), and uncertainty measurement and data referral for the classification of benign and malignant nodules using 3D CT images. To quantify the uncertainty, we employed three approaches; Monte Carlo Dropout (MCD), Deep Ensemble (DE), and Ensemble Monte Carlo Dropout (EMCD). We evaluated eight different deep learning models consisting of ResNet, DenseNet, and the Inception network family, all of which achieved average F1 scores above 0.832, and the highest average value of 0.845 was obtained using InceptionResNetV2. Furthermore, incorporating the UQ demonstrated significant improvement in the overall model performance. Upon evaluation of the uncertainty estimate, MCD outperforms the other UQ models except for the metric, URecall, where DE and EMCD excel, implying that they are better at identifying incorrect predictions with higher uncertainty levels, which is vital in the medical field. Finally, we show that using a threshold for data referral can greatly improve the performance further, increasing the accuracy up to 0.959.
Collapse
|
15
|
Faghani S, Gamble C, Erickson BJ. Uncover This Tech Term: Uncertainty Quantification for Deep Learning. Korean J Radiol 2024; 25:395-398. [PMID: 38528697 PMCID: PMC10973738 DOI: 10.3348/kjr.2024.0108] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [MESH Headings] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/29/2024] [Revised: 02/05/2024] [Accepted: 02/06/2024] [Indexed: 03/27/2024] Open
|
16
|
Park SH, Hwang EJ. Caveats in Using Abnormality/Probability Scores from Artificial Intelligence Algorithms: Neither True Probability nor Level of Trustworthiness. Korean J Radiol 2024; 25:328-330. [PMID: 38528690 PMCID: PMC10973731 DOI: 10.3348/kjr.2024.0144] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [MESH Headings] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2024] [Accepted: 02/08/2024] [Indexed: 03/27/2024] Open
|
17
|
Consagra W, Ning L, Rathi Y. Neural orientation distribution fields for estimation and uncertainty quantification in diffusion MRI. Med Image Anal 2024; 93:103105. [PMID: 38377728 DOI: 10.1016/j.media.2024.103105] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2023] [Revised: 12/13/2023] [Accepted: 02/05/2024] [Indexed: 02/22/2024]
Abstract
Inferring brain connectivity and structure in-vivo requires accurate estimation of the orientation distribution function (ODF), which encodes key local tissue properties. However, estimating the ODF from diffusion MRI (dMRI) signals is a challenging inverse problem due to obstacles such as significant noise, high-dimensional parameter spaces, and sparse angular measurements. In this paper, we address these challenges by proposing a novel deep-learning based methodology for continuous estimation and uncertainty quantification of the spatially varying ODF field. We use a neural field (NF) to parameterize a random series representation of the latent ODFs, implicitly modeling the often ignored but valuable spatial correlation structures in the data, and thereby improving efficiency in sparse and noisy regimes. An analytic approximation to the posterior predictive distribution is derived which can be used to quantify the uncertainty in the ODF estimate at any spatial location, avoiding the need for expensive resampling-based approaches that are typically employed for this purpose. We present empirical evaluations on both synthetic and real in-vivo diffusion data, demonstrating the advantages of our method over existing approaches.
Collapse
|
18
|
Das P, Igoe M, Lacy A, Farthing T, Timsina A, Lanzas C, Lenhart S, Odoi A, Lloyd AL. Modeling county level COVID-19 transmission in the greater St. Louis area: Challenges of uncertainty and identifiability when fitting mechanistic models to time-varying processes. Math Biosci 2024; 371:109181. [PMID: 38537734 DOI: 10.1016/j.mbs.2024.109181] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2023] [Accepted: 03/13/2024] [Indexed: 04/12/2024]
Abstract
We use a compartmental model with a time-varying transmission parameter to describe county level COVID-19 transmission in the greater St. Louis area of Missouri and investigate the challenges in fitting such a model to time-varying processes. We fit this model to synthetic and real confirmed case and hospital discharge data from May to December 2020 and calculate uncertainties in the resulting parameter estimates. We also explore non-identifiability within the estimated parameter set. We find that the death rate of infectious non-hospitalized individuals, the testing parameter and the initial number of exposed individuals are not identifiable based on an investigation of correlation coefficients between pairs of parameter estimates. We also explore how this non-identifiability ties back into uncertainties in the estimated parameters and find that it inflates uncertainty in the estimates of our time-varying transmission parameter. However, we do find that R0 is not highly affected by non-identifiability of its constituent components and the uncertainties associated with the quantity are smaller than those of the estimated parameters. Parameter values estimated from data will always be associated with some uncertainty and our work highlights the importance of conducting these analyses when fitting such models to real data. Exploring identifiability and uncertainty is crucial in revealing how much we can trust the parameter estimates.
Collapse
|
19
|
Roth JP, Bajorath J. Relationship between prediction accuracy and uncertainty in compound potency prediction using deep neural networks and control models. Sci Rep 2024; 14:6536. [PMID: 38503823 PMCID: PMC10950896 DOI: 10.1038/s41598-024-57135-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2023] [Accepted: 03/14/2024] [Indexed: 03/21/2024] Open
Abstract
The assessment of prediction variance or uncertainty contributes to the evaluation of machine learning models. In molecular machine learning, uncertainty quantification is an evolving area of research where currently no standard approaches or general guidelines are available. We have carried out a detailed analysis of deep neural network variants and simple control models for compound potency prediction to study relationships between prediction accuracy and uncertainty. For comparably accurate predictions obtained with models of different complexity, highly variable prediction uncertainties were detected using different metrics. Furthermore, a strong dependence of prediction characteristics and uncertainties on potency levels of test compounds was observed, often leading to over- or under-confident model decisions with respect to the expected variance of predictions. Moreover, neural network models responded very differently to training set modifications. Taken together, our findings indicate that there is only little, if any correlation between compound potency prediction accuracy and uncertainty, especially for deep neural network models, when predictions are assessed on the basis of currently used metrics for uncertainty quantification.
Collapse
|
20
|
Chung SC. Cryo-forum: A framework for orientation recovery with uncertainty measure with the application in cryo-EM image analysis. J Struct Biol 2024; 216:108058. [PMID: 38163450 DOI: 10.1016/j.jsb.2023.108058] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2023] [Revised: 12/14/2023] [Accepted: 12/28/2023] [Indexed: 01/03/2024]
Abstract
In single-particle cryo-electron microscopy (cryo-EM), efficient determination of orientation parameters for particle images poses a significant challenge yet is crucial for reconstructing 3D structures. This task is complicated by the high noise levels in the datasets, which often include outliers, necessitating several time-consuming 2D clean-up processes. Recently, solutions based on deep learning have emerged, offering a more streamlined approach to the traditionally laborious task of orientation estimation. These solutions employ amortized inference, eliminating the need to estimate parameters individually for each image. However, these methods frequently overlook the presence of outliers and may not adequately concentrate on the components used within the network. This paper introduces a novel method using a 10-dimensional feature vector for orientation representation, extracting orientations as unit quaternions with an accompanying uncertainty metric. Furthermore, we propose a unique loss function that considers the pairwise distances between orientations, thereby enhancing the accuracy of our method. Finally, we also comprehensively evaluate the design choices in constructing the encoder network, a topic that has not received sufficient attention in the literature. Our numerical analysis demonstrates that our methodology effectively recovers orientations from 2D cryo-EM images in an end-to-end manner. Notably, the inclusion of uncertainty quantification allows for direct clean-up of the dataset at the 3D level. Lastly, we package our proposed methods into a user-friendly software suite named cryo-forum, designed for easy access by developers.
Collapse
|
21
|
Yin J, Huang Y, Lu C, Liu Z. Uncertainty-based saltwater intrusion prediction using integrated Bayesian machine learning modeling (IBMLM) in a deep aquifer. JOURNAL OF ENVIRONMENTAL MANAGEMENT 2024; 354:120252. [PMID: 38394869 DOI: 10.1016/j.jenvman.2024.120252] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/18/2023] [Revised: 01/25/2024] [Accepted: 01/27/2024] [Indexed: 02/25/2024]
Abstract
Data-driven machine learning approaches are promising to substitute physically based groundwater numerical models and capture input-output relationships for reducing computational burden. But the performance and reliability are strongly influenced by different sources of uncertainty. Conventional researches generally rely on a stand-alone machine learning surrogate approach and fail to account for errors in model outputs resulting from structural deficiencies. To overcome this issue, this study proposes a flexible integrated Bayesian machine learning modeling (IBMLM) method to explicitly quantify uncertainties originating from structures and parameters of machine learning surrogate models. An Expectation-Maximization (EM) algorithm is combined with Bayesian model averaging (BMA) to find out maximum likelihood and construct posterior predictive distribution. Three machine learning approaches representing different model complexity are incorporated in the framework, including artificial neural network (ANN), support vector machine (SVM) and random forest (RF). The proposed IBMLM method is demonstrated in a field-scale real-world "1500-foot" sand aquifer, Baton Rouge, USA, where overexploitation caused serious saltwater intrusion (SWI) issues. This study adds to the understanding of how chloride concentration transport responds to multi-dimensional extraction-injection remediation strategies in a sophisticated saltwater intrusion model. Results show that most IBMLM exhibit r values above 0.98 and NSE values above 0.93, both slightly higher than individual machine learning, confirming that the IBMLM is well established to provide better model predictions than individual machine learning models, while maintaining the advantage of high computing efficiency. The IBMLM is found useful to predict saltwater intrusion without running the physically based numerical simulation model. We conclude that an explicit consideration of machine learning model structure uncertainty along with parameters improves accuracy and reliability of predictions, and also corrects uncertainty bounds. The applicability of the IBMLM framework can be extended in regions where a physical hydrogeologic model is difficult to build due to lack of subsurface information.
Collapse
|
22
|
Gheysen L, Maes L, Caenen A, Segers P, Peirlinck M, Famaey N. Uncertainty quantification of the wall thickness and stiffness in an idealized dissected aorta. J Mech Behav Biomed Mater 2024; 151:106370. [PMID: 38224645 DOI: 10.1016/j.jmbbm.2024.106370] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2023] [Revised: 12/21/2023] [Accepted: 01/01/2024] [Indexed: 01/17/2024]
Abstract
Personalized treatment informed by computational models has the potential to markedly improve the outcome for patients with a type B aortic dissection. However, existing computational models of dissected walls significantly simplify the characteristic false lumen, tears and/or material behavior. Moreover, the patient-specific wall thickness and stiffness cannot be accurately captured non-invasively in clinical practice, which inevitably leads to assumptions in these wall models. It is important to evaluate the impact of the corresponding uncertainty on the predicted wall deformations and stress, which are both key outcome indicators for treatment optimization. Therefore, a physiology-inspired finite element framework was proposed to model the wall deformation and stress of a type B aortic dissection at diastolic and systolic pressure. Based on this framework, 300 finite element analyses, sampled with a Latin hypercube, were performed to assess the global uncertainty, introduced by 4 uncertain wall thickness and stiffness input parameters, on 4 displacement and stress output parameters. The specific impact of each input parameter was estimated using Gaussian process regression, as surrogate model of the finite element framework, and a δ moment-independent analysis. The global uncertainty analysis indicated minor differences between the uncertainty at diastolic and systolic pressure. For all output parameters, the 4th quartile contained the major fraction of the uncertainty. The parameter-specific uncertainty analysis elucidated that the material stiffness and relative thickness of the dissected membrane were the respective main determinants of the wall deformation and stress. The uncertainty analysis provides insight into the effect of uncertain wall thickness and stiffness parameters on the predicted deformation and stress. Moreover, it emphasizes the need for probabilistic rather than deterministic predictions for clinical decision making in aortic dissections.
Collapse
|
23
|
Nicol PB, Miller JW. Model-based dimensionality reduction for single-cell RNA-seq using generalized bilinear models. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2023.04.21.537881. [PMID: 37162914 PMCID: PMC10168202 DOI: 10.1101/2023.04.21.537881] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/11/2023]
Abstract
Dimensionality reduction is a critical step in the analysis of single-cell RNA-seq (scRNA-seq) data. The standard approach is to apply a transformation to the count matrix followed by principal components analysis (PCA). However, this approach can induce spurious heterogeneity and mask true biological variability. An alternative approach is to directly model the counts, but existing methods tend to be computationally intractable on large datasets and do not quantify uncertainty in the low-dimensional representation. To address these problems, we develop scGBM, a novel method for model-based dimensionality reduction of scRNA-seq data using a Poisson bilinear model. We introduce a fast estimation algorithm to fit the model using iteratively reweighted singular value decompositions, enabling the method to scale to datasets with millions of cells. Furthermore, scGBM quantifies the uncertainty in each cell's latent position and leverages these uncertainties to assess the confidence associated with a given cell clustering. On real and simulated single-cell data, we find that scGBM produces low-dimensional embeddings that better capture relevant biological information while removing unwanted variation.
Collapse
|
24
|
Gholami H, Mohammadifar A, Behrooz RD, Kaskaoutis DG, Li Y, Song Y. Intrinsic and extrinsic techniques for quantification uncertainty of an interpretable GRU deep learning model used to predict atmospheric total suspended particulates (TSP) in Zabol, Iran during the dusty period of 120-days wind. ENVIRONMENTAL POLLUTION (BARKING, ESSEX : 1987) 2024; 342:123082. [PMID: 38061429 DOI: 10.1016/j.envpol.2023.123082] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Revised: 11/11/2023] [Accepted: 11/30/2023] [Indexed: 12/17/2023]
Abstract
Total suspended particulates (TSP), as a key pollutant, is a serious threat for air quality, climate, ecosystems and human health. Therefore, measurements, prediction and forecasting of TSP concentrations are necessary to mitigate their negative effects. This study applies the gated recurrent unit (GRU) deep learning model to predict TSP concentrations in Zabol, Iran, during the dust period of the 120-day wind (3 June - 4 October 2014). Three uncertainty quantification (UQ) techniques consisting of the blackbox metamodel, heteroscedastic regression and infinitesimal jackknife were applied to quantify the uncertainty associated with GRU model. Permutation feature importance measure (PFIM), based on the game theory, was employed for the interpretability of the predictive model's outputs. A total of 80 TSP samples were collected and were randomly divided as training (70%) and validation (30%) datasets, while eight variables were used in the TSP prediction model. Our findings showed that GRU performed very well for TSP prediction (with r and Nash Sutcliffe coefficient (NSC) values above 0.99 for both datasets, and RMSE of 57 μg m-3 and 73 μg m-3 for training and validation datasets, respectively). Among the three UQ techniques, the infinitesimal jackknife was the most accurate one, while all the observed and predicted TSP values fell within the continence limitation estimated by the model. PFIM plots showed that wind speed and air humidity were the most and least important variables, respectively, impacting the predictive model's outputs. This is the first attempt of using an interpretable DL model for TSP prediction modelling, recommending that future research should involve aspects of uncertainty and interpretability of the predictive models. Overall, UQ and interpretability techniques have a key role in reducing the impact of uncertainties during optimization and decision making, resulting in better understanding of sophisticated mechanisms related to the predictive model.
Collapse
|
25
|
de Sousa AA, Machado MR. Experimental vibration dataset collected of a beam reinforced with masses under different health conditions. Data Brief 2024; 52:110043. [PMID: 38299099 PMCID: PMC10828560 DOI: 10.1016/j.dib.2024.110043] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2023] [Revised: 12/19/2023] [Accepted: 01/04/2024] [Indexed: 02/02/2024] Open
Abstract
Vibration signals extracted from structures across diverse health conditions have become indispensable for monitoring structural integrity. These datasets represent a resource for real-time condition monitoring, enabling the precise detection and diagnosis of system anomalies. This paper aims to enrich the scientific community's database on structural dynamics and experimental methodologies pertinent to system modelling. Leveraging experimental measurements obtained from mass-reinforced beams, these datasets validate numerical models, refine identification techniques, quantify uncertainties, and continuously foster machine learning algorithms' evolution to monitor structural integrity. Furthermore, the beam dataset is data-driven and can be used to develop and test innovative structural health monitoring strategies, specifically identifying damages and anomalies within intricate structural frameworks. Supplemental datasets like Mass-position and damage index introduce parametric uncertainty into experimental and damage identification metrics. Thereby offering valuable insights to elevate the efficacy of monitoring and control techniques. These comprehensive tests also encapsulate paramedic uncertainty, providing robust support for applications in uncertainty quantification, stochastic modelling, and supervised and unsupervised machine learning methodologies.
Collapse
|