1
|
Karimi Mamaghan AM, Dittadi A, Bauer S, Johansson KH, Quinzan F. Diffusion-Based Causal Representation Learning. ENTROPY (BASEL, SWITZERLAND) 2024; 26:556. [PMID: 39056918 PMCID: PMC11275649 DOI: 10.3390/e26070556] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/30/2024] [Revised: 06/24/2024] [Accepted: 06/25/2024] [Indexed: 07/28/2024]
Abstract
Causal reasoning can be considered a cornerstone of intelligent systems. Having access to an underlying causal graph comes with the promise of cause-effect estimation and the identification of efficient and safe interventions. However, learning causal representations remains a major challenge, due to the complexity of many real-world systems. Previous works on causal representation learning have mostly focused on Variational Auto-Encoders (VAEs). These methods only provide representations from a point estimate, and they are less effective at handling high dimensions. To overcome these problems, we propose a Diffusion-based Causal Representation Learning (DCRL) framework which uses diffusion-based representations for causal discovery in the latent space. DCRL provides access to both single-dimensional and infinite-dimensional latent codes, which encode different levels of information. In a first proof of principle, we investigate the use of DCRL for causal representation learning in a weakly supervised setting. We further demonstrate experimentally that this approach performs comparably well in identifying the latent causal structure and causal variables.
Collapse
Affiliation(s)
| | - Andrea Dittadi
- Helmholtz AI, 85764 Munich, Germany; (A.D.); (S.B.)
- MPI for Intelligent Systems, 72076 Tübingen, Germany
- School of Computation, Information and Technology, TU Munich, 80333 Munich, Germany
| | - Stefan Bauer
- Helmholtz AI, 85764 Munich, Germany; (A.D.); (S.B.)
- School of Computation, Information and Technology, TU Munich, 80333 Munich, Germany
| | - Karl Henrik Johansson
- Division of Decision and Control Systems (DCS), KTH Royal Institute of Technology, 114 28 Stockholm, Sweden;
- Digital Futures, 114 28 Stockholm, Sweden
| | - Francesco Quinzan
- Department of Computer Science, University of Oxford, Oxford OX1 2JD, UK;
| |
Collapse
|
2
|
Post RAJ, van den Heuvel ER, Putter H. Bias of the additive hazard model in the presence of causal effect heterogeneity. LIFETIME DATA ANALYSIS 2024; 30:383-403. [PMID: 38466520 PMCID: PMC10957647 DOI: 10.1007/s10985-024-09616-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/29/2022] [Accepted: 01/10/2024] [Indexed: 03/13/2024]
Abstract
Hazard ratios are prone to selection bias, compromising their use as causal estimands. On the other hand, if Aalen's additive hazard model applies, the hazard difference has been shown to remain unaffected by the selection of frailty factors over time. Then, in the absence of confounding, observed hazard differences are equal in expectation to the causal hazard differences. However, in the presence of effect (on the hazard) heterogeneity, the observed hazard difference is also affected by selection of survivors. In this work, we formalize how the observed hazard difference (from a randomized controlled trial) evolves by selecting favourable levels of effect modifiers in the exposed group and thus deviates from the causal effect of interest. Such selection may result in a non-linear integrated hazard difference curve even when the individual causal effects are time-invariant. Therefore, a homogeneous time-varying causal additive effect on the hazard cannot be distinguished from a time-invariant but heterogeneous causal effect. We illustrate this causal issue by studying the effect of chemotherapy on the survival time of patients suffering from carcinoma of the oropharynx using data from a clinical trial. The hazard difference can thus not be used as an appropriate measure of the causal effect without making untestable assumptions.
Collapse
Affiliation(s)
- Richard A J Post
- Department of Mathematics and Computer Science, Eindhoven University of Technology, Eindhoven, The Netherlands.
| | - Edwin R van den Heuvel
- Department of Mathematics and Computer Science, Eindhoven University of Technology, Eindhoven, The Netherlands
| | - Hein Putter
- Department of Biomedical Data Sciences, Leiden University Medical Center, Leiden, The Netherlands
- Mathematical Institute, Leiden University, Leiden, The Netherlands
| |
Collapse
|
3
|
Post RAJ, van den Heuvel ER, Putter H. The built-in selection bias of hazard ratios formalized using structural causal models. LIFETIME DATA ANALYSIS 2024; 30:404-438. [PMID: 38358572 PMCID: PMC11300553 DOI: 10.1007/s10985-024-09617-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/29/2022] [Accepted: 01/10/2024] [Indexed: 02/16/2024]
Abstract
It is known that the hazard ratio lacks a useful causal interpretation. Even for data from a randomized controlled trial, the hazard ratio suffers from so-called built-in selection bias as, over time, the individuals at risk among the exposed and unexposed are no longer exchangeable. In this paper, we formalize how the expectation of the observed hazard ratio evolves and deviates from the causal effect of interest in the presence of heterogeneity of the hazard rate of unexposed individuals (frailty) and heterogeneity in effect (individual modification). For the case of effect heterogeneity, we define the causal hazard ratio. We show that the expected observed hazard ratio equals the ratio of expectations of the latent variables (frailty and modifier) conditionally on survival in the world with and without exposure, respectively. Examples with gamma, inverse Gaussian and compound Poisson distributed frailty and categorical (harming, beneficial or neutral) distributed effect modifiers are presented for illustration. This set of examples shows that an observed hazard ratio with a particular value can arise for all values of the causal hazard ratio. Therefore, the hazard ratio cannot be used as a measure of the causal effect without making untestable assumptions, stressing the importance of using more appropriate estimands, such as contrasts of the survival probabilities.
Collapse
Affiliation(s)
- Richard A J Post
- Department of Mathematics and Computer Science, Eindhoven University of Technology, Eindhoven, The Netherlands.
| | - Edwin R van den Heuvel
- Department of Mathematics and Computer Science, Eindhoven University of Technology, Eindhoven, The Netherlands
| | - Hein Putter
- Department of Biomedical Data Sciences, Leiden University Medical Center, Leiden, The Netherlands
- Mathematical Institute, Leiden University, Leiden, The Netherlands
| |
Collapse
|
4
|
Hammond L, Fox J, Everitt T, Carey R, Abate A, Wooldridge M. Reasoning about Causality in Games. ARTIF INTELL 2023. [DOI: 10.1016/j.artint.2023.103919] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/08/2023]
|
5
|
Vonk MC, Malekovic N, Bäck T, Kononova AV. Disentangling causality: assumptions in causal discovery and inference. Artif Intell Rev 2023. [DOI: 10.1007/s10462-023-10411-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/01/2023]
Abstract
AbstractCausality has been a burgeoning field of research leading to the point where the literature abounds with different components addressing distinct parts of causality. For researchers, it has been increasingly difficult to discern the assumptions they have to abide by in order to glean sound conclusions from causal concepts or methods. This paper aims to disambiguate the different causal concepts that have emerged in causal inference and causal discovery from observational data by attributing them to different levels of Pearl’s Causal Hierarchy. We will provide the reader with a comprehensive arrangement of assumptions necessary to engage in causal reasoning at the desired level of the hierarchy. Therefore, the assumptions underlying each of these causal concepts will be emphasized and their concomitant graphical components will be examined. We show which assumptions are necessary to bridge the gaps between causal discovery, causal identification and causal inference from a parametric and a non-parametric perspective. Finally, this paper points to further research areas related to the strong assumptions that researchers have glibly adopted to take part in causal discovery, causal identification and causal inference.
Collapse
|
6
|
Vilasini V, Colbeck R. Impossibility of Superluminal Signaling in Minkowski Spacetime Does Not Rule Out Causal Loops. PHYSICAL REVIEW LETTERS 2022; 129:110401. [PMID: 36154414 DOI: 10.1103/physrevlett.129.110401] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/06/2022] [Revised: 05/20/2022] [Accepted: 06/24/2022] [Indexed: 06/16/2023]
Abstract
Causality is fundamental to science, but it appears in several different forms. One is relativistic causality, which is tied to a spacetime structure and forbids signaling outside the future. A second is an operational notion of causation that considers the flow of information between physical systems and interventions on them. In [V. Vilasini and R. Colbeck, General framework for cyclic and fine-tuned causal models and their compatibility with space-time, Phys. Rev. A 106, 032204 (2022).PLRAAN2469-992610.1103/PhysRevA.106.032204], we propose a framework for characterizing when a causal model can coexist with relativistic principles such as no superluminal signaling, while allowing for cyclic and nonclassical causal influences and the possibility of causation without signaling. In a theory without superluminal causation, both superluminal signaling and causal loops are not possible in Minkowski spacetime. Here we demonstrate that if we only forbid superluminal signaling, superluminal causation remains possible and show the mathematical possibility of causal loops that can be embedded in a Minkowski spacetime without leading to superluminal signaling. The existence of such loops in the given spacetime could in principle be operationally verified using interventions. This establishes that the physical principle of no superluminal signaling is not by itself sufficient to rule out causal loops between Minkowski spacetime events. Interestingly, the conditions required to rule out causal loops in a spacetime depend on the dimension. Whether such loops are possible in three spatial dimensions remains an important open question.
Collapse
Affiliation(s)
- V Vilasini
- Institute for Theoretical Physics, ETH Zurich, 8093 Zürich, Switzerland
- Department of Mathematics, University of York, Heslington, York YO10 5DD, United Kingdom
| | - Roger Colbeck
- Department of Mathematics, University of York, Heslington, York YO10 5DD, United Kingdom
| |
Collapse
|
7
|
Squires C, Uhler C. Causal Structure Learning: A Combinatorial Perspective. FOUNDATIONS OF COMPUTATIONAL MATHEMATICS (NEW YORK, N.Y.) 2022; 23:1-35. [PMID: 35935470 PMCID: PMC9342837 DOI: 10.1007/s10208-022-09581-9] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/28/2022] [Accepted: 06/08/2022] [Indexed: 05/29/2023]
Abstract
In this review, we discuss approaches for learning causal structure from data, also called causal discovery. In particular, we focus on approaches for learning directed acyclic graphs and various generalizations which allow for some variables to be unobserved in the available data. We devote special attention to two fundamental combinatorial aspects of causal structure learning. First, we discuss the structure of the search space over causal graphs. Second, we discuss the structure of equivalence classes over causal graphs, i.e., sets of graphs which represent what can be learned from observational data alone, and how these equivalence classes can be refined by adding interventional data.
Collapse
Affiliation(s)
| | - Caroline Uhler
- Broad Institute and Massachusetts Institute of Technology, Cambridge, MA 02139 USA
| |
Collapse
|
8
|
Sanchez P, Voisey JP, Xia T, Watson HI, O’Neil AQ, Tsaftaris SA. Causal machine learning for healthcare and precision medicine. ROYAL SOCIETY OPEN SCIENCE 2022; 9:220638. [PMID: 35950198 PMCID: PMC9346354 DOI: 10.1098/rsos.220638] [Citation(s) in RCA: 25] [Impact Index Per Article: 12.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/13/2022] [Accepted: 07/15/2022] [Indexed: 06/15/2023]
Abstract
Causal machine learning (CML) has experienced increasing popularity in healthcare. Beyond the inherent capabilities of adding domain knowledge into learning systems, CML provides a complete toolset for investigating how a system would react to an intervention (e.g. outcome given a treatment). Quantifying effects of interventions allows actionable decisions to be made while maintaining robustness in the presence of confounders. Here, we explore how causal inference can be incorporated into different aspects of clinical decision support systems by using recent advances in machine learning. Throughout this paper, we use Alzheimer's disease to create examples for illustrating how CML can be advantageous in clinical scenarios. Furthermore, we discuss important challenges present in healthcare applications such as processing high-dimensional and unstructured data, generalization to out-of-distribution samples and temporal relationships, that despite the great effort from the research community remain to be solved. Finally, we review lines of research within causal representation learning, causal discovery and causal reasoning which offer the potential towards addressing the aforementioned challenges.
Collapse
Affiliation(s)
- Pedro Sanchez
- School of Engineering, University of Edinburgh, Edinburgh, UK
| | - Jeremy P. Voisey
- AI Research, Canon Medical Research Europe, Edinburgh, Lothian, UK
| | - Tian Xia
- School of Engineering, University of Edinburgh, Edinburgh, UK
| | - Hannah I. Watson
- AI Research, Canon Medical Research Europe, Edinburgh, Lothian, UK
| | - Alison Q. O’Neil
- School of Engineering, University of Edinburgh, Edinburgh, UK
- AI Research, Canon Medical Research Europe, Edinburgh, Lothian, UK
| | | |
Collapse
|
9
|
|