A significant hurdle in discovering causal structure from observational data is the potential presence of unobserved common causes, often called latent variables or hidden confounders. Standard algorithms like the basic PC algorithm operate under the assumption of causal sufficiency, meaning that all relevant common causes of the measured variables are themselves measured. When this assumption is violated, which is common in real-world systems, these algorithms can produce incorrect causal graphs, leading to spurious connections or missed causal links. This section details methods designed to perform causal discovery even when causal sufficiency cannot be guaranteed.
Consider a simple true causal structure where an unobserved variable U confounds the relationship between measured variables X and Y: X←U→Y. If U is not measured, an algorithm assuming causal sufficiency might incorrectly infer a direct edge between X and Y (e.g., X→Y or Y→X) based purely on the observed statistical dependence (X⊥Y). Similarly, if X and Y both cause Z, but are also confounded by U (X←U→Y, X→Z, Y→Z), standard algorithms might struggle to orient the edges involving Z correctly, especially if conditioning on Z induces a dependency between X and Y (collider effect).
Managing latent variables in discovery involves detecting their potential influence and representing the resulting structural uncertainty, rather than necessarily identifying the exact latent variables themselves.
The Fast Causal Inference (FCI) algorithm is a cornerstone constraint-based method explicitly designed to handle latent variables. Unlike PC, FCI does not assume causal sufficiency. It aims to infer properties of the causal structure that hold true across all possible underlying Directed Acyclic Graphs (DAGs) compatible with the observed conditional independencies, including those with latent confounders.
Instead of outputting a single DAG or a Markov equivalence class of DAGs (like PC), FCI outputs a Partial Ancestral Graph (PAG). A PAG represents a class of DAGs, including those with latent variables, that are consistent with the data. PAGs use a richer set of edge markings to encode the uncertainty introduced by potential hidden confounders:
Example PAG with different edge types. X→Y denotes a definite causal ancestry. Y↔Z (represented visually with a double arrow) indicates definite confounding. Z o→W means W is not an ancestor of Z, but Z could cause W or they could be confounded. X o−o W represents uncertainty about whether X causes W, W causes X, or they share a common confounder. The dashed node L suggests a potential latent variable structure consistent with Y↔Z.
FCI operates similarly to PC by iteratively testing conditional independencies. However, its rules for determining adjacencies and orientations are more complex to account for potential latent variables. It typically involves:
Extensions like RFCI (Really Fast Causal Inference) and FCI+ aim to improve the computational efficiency and, in some cases, the accuracy of FCI, particularly for high-dimensional datasets.
While powerful, FCI comes with important considerations:
Standard score-based algorithms, such as Greedy Equivalence Search (GES), typically assume causal sufficiency. They search for the Markov equivalence class of DAGs that best fits the data according to a scoring criterion (e.g., BIC, BDeu). When latent confounders are present, the true causal structure might not be representable by any DAG over the observed variables only, potentially leading GES to converge to an incorrect structure.
While research exists on extending score-based methods to handle latent variables (e.g., by explicitly modeling latent variables or using scores robust to certain types of confounding), these approaches are often more complex and less generally applicable than FCI. Methods like Linear Non-Gaussian Acyclic Models (LiNGAM) can sometimes distinguish direct causation from confounding under specific assumptions (non-Gaussianity of error terms), but they don't provide a general solution for arbitrary latent variable structures.
Python libraries such as causal-learn
provide implementations of FCI, RFCI, and related algorithms. When using these tools:
alpha
) for the CI tests, as the resulting PAG can be sensitive to this parameter.Effectively managing latent variables is fundamental for trustworthy causal discovery from observational data. The assumption of causal sufficiency is often unrealistic. Constraint-based algorithms like FCI provide a principled way to infer causal structures while acknowledging the possible presence of hidden confounders, yielding Partial Ancestral Graphs (PAGs) that represent the inherent uncertainty. While standard score-based methods struggle with latents, and specialized methods exist for specific cases, FCI remains a standard approach for general discovery under potential confounding. Interpreting the resulting PAGs requires care, recognizing that they represent equivalence classes of causal models rather than a single definitive structure.
© 2025 ApX Machine Learning