{"id":3104,"date":"2025-06-27T09:39:01","date_gmt":"2025-06-27T09:39:01","guid":{"rendered":"https:\/\/uplatz.com\/blog\/?p=3104"},"modified":"2025-06-27T09:39:01","modified_gmt":"2025-06-27T09:39:01","slug":"causal-graph-learning-in-observational-data","status":"publish","type":"post","link":"https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/","title":{"rendered":"Causal Graph Learning in Observational Data"},"content":{"rendered":"<h1><b>Causal Graph Learning in Observational Data<\/b><\/h1>\n<h2><b>I. Introduction to Causal Graph Learning in Observational Data<\/b><\/h2>\n<p><span style=\"font-weight: 400;\">Causal graph learning represents a pivotal subfield within machine learning, shifting the focus from mere predictive modeling to the identification of genuine cause-and-effect relationships between variables. This distinction is paramount for developing artificial intelligence systems that transcend simple prediction, enabling them to comprehend the underlying mechanisms of a system and make more informed, robust decisions.<\/span><span style=\"font-weight: 400;\">1<\/span><span style=\"font-weight: 400;\"> Unlike traditional statistical methods that often highlight correlations, causal learning aims to uncover the directional influences that drive observed phenomena.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Central to this endeavor are causal graph models, frequently depicted as Directed Acyclic Graphs (DAGs). These graphical representations utilize nodes to symbolize variables and directed edges to denote assumed causal influences. Such models are indispensable tools for explicitly mapping out hypothesized causal structures, which is crucial for predicting the outcomes of interventions and fostering a deeper understanding of complex systems.<\/span><span style=\"font-weight: 400;\">1<\/span><span style=\"font-weight: 400;\"> For instance, in health research, the primary objective often involves identifying and quantifying risk factors that exert a causal effect on health and social outcomes.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> This foundational understanding underscores that causal inference seeks to move beyond statistical associations to actionable insights, a prerequisite for effective intervention and scientific advancement.<\/span><\/p>\n<p>&nbsp;<\/p>\n<h3><b>The Fundamental Problem of Causal Inference<\/b><\/h3>\n<p><span style=\"font-weight: 400;\">The core conceptual hurdle in causal inference is widely recognized as the &#8220;fundamental problem of causal inference&#8221;: the inherent impossibility of directly observing counterfactuals. For any given individual, only one of two potential outcomes can ever be observed\u2014either what transpired under a specific exposure (e.g., a medical treatment) or what would have occurred under a different exposure (e.g., no treatment)\u2014but never both simultaneously.<\/span><span style=\"font-weight: 400;\">3<\/span><span style=\"font-weight: 400;\"> This inherent limitation means that the individual causal effect, often denoted as<\/span><\/p>\n<p><span style=\"font-weight: 400;\">\u03c4i\u200b, cannot be empirically measured directly.<\/span><span style=\"font-weight: 400;\">3<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Consequently, causal inference in practice relies on estimating average causal effects across groups of individuals. This involves comparing the risk or outcome in an exposed group to that in an unexposed group.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> This task, however, remains distinct from comparing an individual to their own unobserved counterfactual state, necessitating meticulous methodological considerations. The pervasive nature of this impossibility transforms causal inference from a straightforward data-driven task into a rigorous exercise in assumption validation and sensitivity analysis. The entire framework of causal inference from observational data is constructed upon the imputation or approximation of these unobserved states, which inherently necessitates the introduction of strong, often untestable, assumptions.<\/span><span style=\"font-weight: 400;\">6<\/span><span style=\"font-weight: 400;\"> This emphasizes that causal inference is less about directly discovering an inherent truth from data and more about constructing a plausible causal narrative under specific, often stringent, assumptions, thereby shifting the focus from simple data analysis to rigorous assumption validation and sensitivity analysis.<\/span><\/p>\n<h3><b>Why Observational Data Poses Unique Challenges<\/b><\/h3>\n<p><span style=\"font-weight: 400;\">Observational data, by its very nature, is collected without deliberate interventions or randomized assignments, rendering it intrinsically susceptible to various biases that can obscure or distort true causal relationships.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> These biases include confounding, selection bias, and measurement bias.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> Unlike randomized controlled trials (RCTs), which are considered the &#8220;gold standard&#8221; for causal inference due to their ability to balance confounders across groups through random assignment, observational studies inherently lack this balance.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> The absence of randomization implies that observed associations may stem from common causes (confounders) rather than direct causal links, making the inference of causation significantly more complex and requiring advanced techniques to mitigate bias.<\/span><span style=\"font-weight: 400;\">7<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The impracticality, expense, or ethical constraints associated with conducting RCTs in many real-world scenarios often make observational data the only available option.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> This forces researchers to utilize observational data despite its inherent susceptibility to biases. This situation highlights an inherent, often unavoidable, trade-off between the &#8220;purity&#8221; of causal inference, which RCTs achieve through design-based bias minimization, and the &#8220;practicality&#8221; of data collection, where observational data is readily available but prone to biases. The decision to use observational data is thus frequently a pragmatic necessity rather than a choice of convenience. This necessity, in turn, drives the imperative for developing and applying sophisticated bias mitigation techniques and demanding careful assumption validation.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> The most effective methods in practice are often those that can robustly manage the specific biases present in the available observational data, even if it means accepting certain compromises on theoretical ideals.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The table below delineates the fundamental differences between correlation and causation, underscoring why causal graph learning is essential for moving beyond mere statistical associations to understanding underlying mechanisms and enabling effective interventions.<\/span><\/p>\n<table>\n<tbody>\n<tr>\n<td><span style=\"font-weight: 400;\">Feature<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Correlation<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Causation<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Definition<\/b><\/td>\n<td><span style=\"font-weight: 400;\">Statistical association between variables where changes in one are related to changes in another.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">A relationship where a change in one variable directly leads to a change in another, implying a cause-and-effect link.<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Data Source<\/b><\/td>\n<td><span style=\"font-weight: 400;\">Can be observed in any type of data (observational, experimental).<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Requires experimental\/interventional data or observational data analyzed under strong, explicit assumptions.<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Key Challenge<\/b><\/td>\n<td><span style=\"font-weight: 400;\">Spurious associations due to confounding or other biases.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">The fundamental problem of causal inference (unobservable counterfactuals) and the presence of various biases.<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Primary Goal<\/b><\/td>\n<td><span style=\"font-weight: 400;\">Prediction, description, pattern recognition.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Understanding underlying mechanisms, predicting outcomes of interventions, informing policy.<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Example<\/b><\/td>\n<td><span style=\"font-weight: 400;\">Falling barometer and storm.<\/span><span style=\"font-weight: 400;\">9<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Medicine and patient survival.<\/span><span style=\"font-weight: 400;\">10<\/span><\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p>&nbsp;<\/p>\n<h2><b>II. Challenges and Biases in Causal Inference from Observational Data<\/b><\/h2>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400;\">Drawing valid causal inferences from observational data is fraught with challenges due to inherent biases. A comprehensive understanding of these biases and the stringent assumptions required for their mitigation is critical for robust causal analysis.<\/span><\/p>\n<p>&nbsp;<\/p>\n<h3><b>Confounding Bias<\/b><\/h3>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400;\">Confounding bias arises when an uncontrolled common cause, termed a &#8220;confounder,&#8221; influences both the exposure (treatment variable) and the outcome variable. This simultaneous influence creates a spurious, non-causal association between the exposure and outcome, leading to misleading conclusions.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> In the context of a Directed Acyclic Graph (DAG), confounding is visually represented by an &#8220;open back-door path&#8221; between the exposure (A) and the outcome (Y) that remains unblocked by conditioning on other variables.<\/span><span style=\"font-weight: 400;\">5<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Confounders can manifest in various forms. &#8220;Observed confounders&#8221; are those for which measurements are available in the study data, allowing for statistical adjustment.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> However, a significant challenge arises from &#8220;unmeasured&#8221; or &#8220;unobserved confounders,&#8221; which lead to &#8220;residual confounding&#8221; even after adjusting for all known variables.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> This situation directly violates the crucial ignorability assumption, making unbiased causal estimation exceedingly difficult.<\/span><span style=\"font-weight: 400;\">3<\/span><span style=\"font-weight: 400;\"> Furthermore, confounders can be &#8220;time-varying,&#8221; changing over time for an individual, or &#8220;time-invariant,&#8221; remaining static (e.g., age).<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> Handling time-varying confounders that are themselves affected by prior exposure often necessitates specialized methods, such as marginal structural models.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> While traditional statistical adjustments, including regression models and propensity score methods, aim to control for confounding, they frequently rely on the strong and often unrealistic assumption that all potential confounders are accurately measured and included.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> Fixed-effects regression models offer a partial solution by accounting for unobserved time-invariant confounders.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> The pervasive nature of confounding bias directly undermines the internal validity of causal claims, making it impossible to distinguish genuine cause-effect relationships from mere statistical associations.<\/span><\/p>\n<p>&nbsp;<\/p>\n<h3><b>Selection Bias and Collider Bias<\/b><\/h3>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400;\">Selection bias, within the framework of causal inference, is specifically understood as a type of &#8220;collider bias&#8221;.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> A &#8220;collider&#8221; is a variable that serves as a common effect of two or more other variables, graphically represented by two arrows &#8220;colliding&#8221; into it on a DAG.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> Critically, unlike confounders, which introduce bias if<\/span><\/p>\n<p><i><span style=\"font-weight: 400;\">not<\/span><\/i><span style=\"font-weight: 400;\"> conditioned on, colliders can introduce bias if they <\/span><i><span style=\"font-weight: 400;\">are<\/span><\/i><span style=\"font-weight: 400;\"> conditioned on.<\/span><span style=\"font-weight: 400;\">5<\/span><span style=\"font-weight: 400;\"> This conditioning, often occurring through the selection of a study sample based on the collider&#8217;s value, can inadvertently &#8220;open up&#8221; a spurious back-door path between its causes, thereby creating a non-causal association.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">This bias can manifest in various ways, including differential loss to follow-up, non-response bias, or the inappropriate selection of participants.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> A classic example is &#8220;Berkson&#8217;s bias,&#8221; where restricting a sample to hospitalized patients can create spurious negative associations between otherwise unrelated risk factors.<\/span><span style=\"font-weight: 400;\">5<\/span><span style=\"font-weight: 400;\"> Collider bias often proves counter-intuitive for researchers accustomed to simply &#8220;controlling&#8221; for all available variables. This highlights that indiscriminately including covariates in a model can<\/span><\/p>\n<p><i><span style=\"font-weight: 400;\">introduce<\/span><\/i><span style=\"font-weight: 400;\"> bias rather than remove it, potentially leading to what are termed &#8220;garbage-can regressions&#8221;.<\/span><span style=\"font-weight: 400;\">7<\/span><span style=\"font-weight: 400;\"> This underscores the non-mechanistic nature of causal inference and the imperative of utilizing causal graphs and domain knowledge to guide judicious covariate selection.<\/span><span style=\"font-weight: 400;\">3<\/span><\/p>\n<p>&nbsp;<\/p>\n<h3><b>Measurement Bias<\/b><\/h3>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400;\">Measurement bias, also known as measurement error, occurs when the observed values of a variable systematically deviate from its true, unobserved value.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> This can stem from imprecise data collection methods or misreporting.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> Both differential (where error is related to another variable) and non-differential measurement errors can lead to biased causal estimates.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> For instance, measurement error in a mediating variable can result in an underestimation of the indirect effect and an overestimation of the direct effect.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> Furthermore, differential measurement error can inadvertently &#8220;open&#8221; backdoor pathways between exposure and outcome, effectively inducing confounding.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> The potential outcomes framework conceptualizes measurement error as a form of missing data, where the true values remain unobserved.<\/span><span style=\"font-weight: 400;\">6<\/span><span style=\"font-weight: 400;\"> One approach to mitigate measurement bias involves employing &#8220;latent variables,&#8221; which are statistical constructs estimated from the covariation among strongly related observed variables.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> If these observed variables are assessed using multiple methods with different sources of bias, the latent variable approach can help in removing variability attributable to shared biases.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> Measurement bias is a universal challenge in empirical research; even with a perfectly understood causal structure and identified confounders, inaccurate measurement can still lead to flawed causal conclusions, emphasizing the need for robust data collection and methods that account for measurement uncertainty.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The interconnectedness of these biases is a critical aspect of causal inference. For instance, selection bias is fundamentally a type of collider bias <\/span><span style=\"font-weight: 400;\">5<\/span><span style=\"font-weight: 400;\">, and measurement error can, under certain conditions (e.g., differential error), open up backdoor pathways, thereby inducing confounding.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> This implies that a single flawed methodological decision, such as conditioning on a collider, can cascade into multiple forms of bias, creating a complex web of distortions. Directed Acyclic Graphs (DAGs) become indispensable not just for identifying individual biases but for diagnosing the interplay of biases and understanding how a chosen adjustment strategy might inadvertently introduce new ones. They serve as a &#8220;gold standard tool&#8221; for assessing bias likelihood and guiding appropriate adjustment sets.<\/span><span style=\"font-weight: 400;\">13<\/span><span style=\"font-weight: 400;\"> This highlights that causal inference is not a mechanical process of simply &#8220;throwing in control variables&#8221;.<\/span><span style=\"font-weight: 400;\">7<\/span><span style=\"font-weight: 400;\"> Instead, it demands a deep, graph-theoretic understanding of how variables relate to avoid &#8220;garbage-can regressions&#8221; <\/span><span style=\"font-weight: 400;\">7<\/span><span style=\"font-weight: 400;\"> and ensure the validity of causal conclusions. Researchers must be rigorously trained in graphical causal models to effectively navigate the complexities of observational data and make defensible causal claims.<\/span><\/p>\n<p>&nbsp;<\/p>\n<h3><b>Key Identification Assumptions<\/b><\/h3>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400;\">For valid causal inference from observational data, it is imperative to translate real-world causal questions into statistical parameters that can be causally interpreted. This translation hinges on satisfying a set of strong identification assumptions.<\/span><span style=\"font-weight: 400;\">4<\/span><span style=\"font-weight: 400;\"> While these assumptions are generally guaranteed by design in a well-executed randomized controlled trial, they must be meticulously considered and, whenever possible, validated in observational studies.<\/span><\/p>\n<ul>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><b>Exchangeability<\/b><span style=\"font-weight: 400;\">: This assumption posits that, conditional on observed covariates, the treatment groups are comparable and well-balanced with respect to the distribution of both measured and unmeasured confounders.<\/span><span style=\"font-weight: 400;\">4<\/span><span style=\"font-weight: 400;\"> In essence, it implies that the exposed and unexposed groups are &#8220;exchangeable&#8221; in terms of their potential outcomes, as if treatment had been randomly assigned.<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><b>Consistency<\/b><span style=\"font-weight: 400;\">: This assumption requires that the observed treatment levels in the collected data correspond to well-defined versions of the treatment.<\/span><span style=\"font-weight: 400;\">4<\/span><span style=\"font-weight: 400;\"> It ensures that the observed outcome for an individual who received a specific treatment is indeed the outcome that would have been observed if that individual had received that specific treatment.<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><b>Positivity<\/b><span style=\"font-weight: 400;\">: This assumption dictates that there must be a non-zero positive probability of receiving every level of treatment for all individuals in the study population.<\/span><span style=\"font-weight: 400;\">4<\/span><span style=\"font-weight: 400;\"> This ensures that all subgroups of the population have a chance to be exposed or unexposed, allowing for meaningful comparisons. Violation can occur if data for a subpopulation is entirely absent.<\/span><span style=\"font-weight: 400;\">12<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><b>Causal Sufficiency (Markovianity)<\/b><span style=\"font-weight: 400;\">: This is a strong assumption, particularly for constraint-based algorithms like PC. It states that the set of observed variables includes all common causes for any pair of variables, effectively implying the absence of unmeasured confounders.<\/span><span style=\"font-weight: 400;\">14<\/span><span style=\"font-weight: 400;\"> This assumption is frequently violated in real-world observational settings.<\/span><span style=\"font-weight: 400;\">8<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><b>Causal Markov Condition<\/b><span style=\"font-weight: 400;\">: This condition assumes that each variable in the causal graph is conditionally independent of its non-descendants, given its direct causes (parents).<\/span><span style=\"font-weight: 400;\">14<\/span><span style=\"font-weight: 400;\"> It links the causal structure of the graph to the observed conditional independencies in the data.<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><b>Faithfulness<\/b><span style=\"font-weight: 400;\">: This assumption is the converse of the Causal Markov Condition. It implies that every conditional independence observed in the data is <\/span><i><span style=\"font-weight: 400;\">exactly<\/span><\/i><span style=\"font-weight: 400;\"> represented by the causal graph.<\/span><span style=\"font-weight: 400;\">14<\/span><span style=\"font-weight: 400;\"> This means the causal graph captures all conditional independencies without including &#8220;extra&#8221; ones. This assumption is a &#8220;critical bottleneck&#8221; for neural causal discovery methods, as it is frequently violated across reasonable dataset sizes, undermining their performance.<\/span><span style=\"font-weight: 400;\">19<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><b>Acyclicity<\/b><span style=\"font-weight: 400;\">: This assumption requires that the causal graph is a Directed Acyclic Graph (DAG), meaning there are no directed cycles or feedback loops.<\/span><span style=\"font-weight: 400;\">12<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><b>Large Sample Size<\/b><span style=\"font-weight: 400;\">: For statistical tests used in causal discovery algorithms (e.g., conditional independence tests), a sufficiently large sample size is required to accurately detect relationships.<\/span><span style=\"font-weight: 400;\">14<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><b>Appropriate Statistical Test<\/b><span style=\"font-weight: 400;\">: The choice of the statistical test for independence must be suitable for the nature of the data (e.g., continuous, categorical, mixed) to ensure reliable detection of conditional independence relationships.<\/span><span style=\"font-weight: 400;\">14<\/span><\/li>\n<\/ul>\n<p><span style=\"font-weight: 400;\">These assumptions form the theoretical bedrock of causal inference. Their validity is paramount, as violations can lead to incorrect or misleading causal conclusions, irrespective of the sophistication of the statistical methods employed. The challenge lies in their untestability from data alone, often requiring substantial domain knowledge and careful reasoning. The pervasive nature of assumption violations, particularly concerning unmeasured confounding and faithfulness, reveals a fundamental &#8220;fragility&#8221; in drawing definitive causal conclusions from single observational studies. This inherent fragility necessitates the adoption of &#8220;triangulation of evidence&#8221;.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> Triangulation, defined as integrating results from several different approaches\u2014each with different and largely unrelated sources of potential bias\u2014provides a &#8220;stronger basis&#8221; for causal inference.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> It is not about finding a single perfect method, but about using multiple, diverse methods to cross-validate findings, acknowledging that each method has inherent weaknesses. This proactive approach, which can involve pre-registering triangulation strategies, enhances the robustness of findings. This suggests a paradigm shift from seeking a single &#8220;true&#8221; causal graph to understanding the robustness and sensitivity of causal claims across different modeling assumptions and methods. A comprehensive causal analysis should therefore include rigorous sensitivity analyses to assumption violations and, ideally, integrate findings from diverse methodological paradigms to build a more compelling body of evidence.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The following table provides a concise overview of the biases and identification assumptions discussed, highlighting their mechanisms, impacts, and associated challenges.<\/span><\/p>\n<table>\n<tbody>\n<tr>\n<td><span style=\"font-weight: 400;\">Bias\/Assumption<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Definition<\/span><\/td>\n<td><span style=\"font-weight: 400;\">DAG Representation\/Mechanism<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Impact on Inference<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Key Challenge\/Limitation<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Relevant Sources<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Confounding Bias<\/b><\/td>\n<td><span style=\"font-weight: 400;\">An uncontrolled common cause influencing both exposure and outcome, creating a spurious association.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Open back-door path.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Leads to spurious associations and biased estimates.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Unmeasured confounders (residual confounding); often unrealistic assumption that all are measured.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">2<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Selection Bias (incl. Collider Bias)<\/b><\/td>\n<td><span style=\"font-weight: 400;\">Occurs when conditioning on a common effect (collider) of two variables, creating a spurious association between them.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Conditioning on a common effect (collider) opens a path.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Can introduce bias if conditioned on; counter-intuitive.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Often introduced by improper control; difficult to identify without domain knowledge.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">2<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Measurement Bias<\/b><\/td>\n<td><span style=\"font-weight: 400;\">Observed values deviate systematically from true values due to imprecise methods or misreporting.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">True (unobserved) and measured values as distinct variables; differential error can open back-door paths.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Distorts true relationships, can induce confounding.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Difficult to quantify and correct for; requires robust data collection.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">2<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Exchangeability<\/b><\/td>\n<td><span style=\"font-weight: 400;\">Treatment groups are comparable regarding measured and unmeasured confounders.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Assumed comparability of groups.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Enables causal interpretation of statistical parameters.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Often unrealistic in observational data; difficult to verify.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">4<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Consistency<\/b><\/td>\n<td><span style=\"font-weight: 400;\">Observed treatment levels correspond to well-defined versions of treatment.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Well-defined treatment assignments.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Ensures observed outcome is the true potential outcome under treatment.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Requires precise definition of interventions.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">4<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Positivity<\/b><\/td>\n<td><span style=\"font-weight: 400;\">Non-zero probability of receiving every level of treatment for all individuals.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">All subgroups have a chance to be exposed\/unexposed.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Ensures meaningful comparisons across groups.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Violation if data for a subpopulation is absent.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">4<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Causal Sufficiency<\/b><\/td>\n<td><span style=\"font-weight: 400;\">All common causes for any pair of variables are observed.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">No unmeasured confounders.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Guarantees identifiability of causal structure.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Often unrealistic in real-world data; limits applicability of some algorithms.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">14<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Causal Markov Condition<\/b><\/td>\n<td><span style=\"font-weight: 400;\">Each variable is independent of its non-descendants, given its direct causes (parents).<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Implied by d-separation patterns in the graph.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Links causal structure to observed conditional independencies.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Assumed property of the data-generating process.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">14<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Faithfulness<\/b><\/td>\n<td><span style=\"font-weight: 400;\">All observed conditional independencies are exactly represented by the causal graph.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Graph captures all conditional independencies without &#8220;extra&#8221; ones.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Essential for accurate structure recovery.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Can be violated in practice, especially for neural methods; critical bottleneck.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">14<\/span><\/td>\n<\/tr>\n<tr>\n<td><b>Acyclicity<\/b><\/td>\n<td><span style=\"font-weight: 400;\">The causal graph contains no directed cycles or feedback loops.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Directed Acyclic Graph (DAG).<\/span><\/td>\n<td><span style=\"font-weight: 400;\">Ensures a well-defined causal ordering.<\/span><\/td>\n<td><span style=\"font-weight: 400;\">May not hold in systems with feedback loops (requires specialized methods).<\/span><\/td>\n<td><span style=\"font-weight: 400;\">12<\/span><\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p>&nbsp;<\/p>\n<h2><b>III. Causal Discovery Algorithms for Observational Data<\/b><\/h2>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400;\">The complexities and biases inherent in observational data necessitate sophisticated algorithmic approaches for causal graph learning. These methods broadly fall into constraint-based, score-based, and hybrid categories, each with distinct principles, assumptions, and limitations.<\/span><\/p>\n<p>&nbsp;<\/p>\n<h3><b>Constraint-Based Methods<\/b><\/h3>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400;\">Constraint-based algorithms infer causal structure by identifying conditional independence (CI) relationships within the observed data. They typically commence with a fully connected undirected graph and iteratively remove edges that are found to be conditionally independent given a subset of other variables. Following this &#8220;skeleton discovery&#8221; phase, a set of orientation rules is applied to determine the direction of causal links, such as identifying v-structures (where two variables cause a third, but are not directly connected themselves) and avoiding cycles.<\/span><span style=\"font-weight: 400;\">15<\/span><span style=\"font-weight: 400;\"> The ultimate objective is to recover the Markov equivalence class of the true causal graph, which represents a set of graphs that imply the same conditional independencies.<\/span><span style=\"font-weight: 400;\">15<\/span><\/p>\n<p>&nbsp;<\/p>\n<h4><b>PC Algorithm<\/b><\/h4>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400;\">The PC (Peter-Clark) algorithm stands as a foundational constraint-based method. It initiates with a complete undirected graph and systematically performs conditional independence tests. If two variables (X, Y) are found to be conditionally independent given a subset of other variables (S), the edge between X and Y is removed. This process is iterative, progressively increasing the size of the conditioning set (n) in each step.<\/span><span style=\"font-weight: 400;\">14<\/span><span style=\"font-weight: 400;\"> After identifying the &#8220;skeleton&#8221; (the undirected graph representing adjacencies), the algorithm applies a series of orientation rules. These rules include identifying &#8220;v-structures&#8221; (e.g., X \u2192 Z \u2190 Y where X and Y are not adjacent) and applying Meek&#8217;s rules to orient remaining edges and prevent the formation of cycles.<\/span><span style=\"font-weight: 400;\">14<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The PC algorithm relies on several strong assumptions for its correctness. These include <\/span><b>Causal Sufficiency<\/b><span style=\"font-weight: 400;\"> (the absence of latent confounders, meaning all common causes are observed), the <\/span><b>Causal Markov Condition<\/b><span style=\"font-weight: 400;\">, <\/span><b>Faithfulness<\/b><span style=\"font-weight: 400;\"> (all observed conditional independencies are implied by the graph), <\/span><b>Causal Ordering<\/b><span style=\"font-weight: 400;\"> (the existence of a causal order among variables), the availability of a <\/span><b>Large Sample Size<\/b><span style=\"font-weight: 400;\"> for reliable CI tests, and the use of an <\/span><b>Appropriate Statistical Test<\/b><span style=\"font-weight: 400;\"> for independence.<\/span><span style=\"font-weight: 400;\">14<\/span><span style=\"font-weight: 400;\"> To address some limitations of the original PC algorithm, several variants have been developed. These include<\/span><\/p>\n<p><b>PC-Stable<\/b><span style=\"font-weight: 400;\">, designed to mitigate order dependence in test results; <\/span><b>Conservative-PC (CPC)<\/b><span style=\"font-weight: 400;\"> and <\/span><b>PC-Max<\/b><span style=\"font-weight: 400;\">, which employ more cautious strategies for orienting colliders; and <\/span><b>Copula-PC<\/b><span style=\"font-weight: 400;\">, specifically engineered to handle mixed continuous and ordinal datasets by inferring rank correlation.<\/span><span style=\"font-weight: 400;\">23<\/span><span style=\"font-weight: 400;\"> The PC algorithm is a cornerstone of causal discovery, demonstrating the power of conditional independence tests to infer causal structure. Its explicit reliance on strong assumptions, such as causal sufficiency, highlights the ideal, often unattainable, conditions under which a definitive causal graph can be learned, thus setting a benchmark for more robust algorithms.<\/span><\/p>\n<p>&nbsp;<\/p>\n<h4><b>FCI Algorithm (Fast Causal Inference)<\/b><\/h4>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400;\">The Fast Causal Inference (FCI) algorithm is a significant extension of the PC algorithm, specifically designed to infer causal relationships from observational data even in the presence of <\/span><b>hidden variables (latent confounders)<\/b><span style=\"font-weight: 400;\"> and <\/span><b>selection bias<\/b><span style=\"font-weight: 400;\">.<\/span><span style=\"font-weight: 400;\">16<\/span><span style=\"font-weight: 400;\"> Unlike PC, which outputs a Completed Partially Directed Acyclic Graph (CPDAG) representing a Markov equivalence class under causal sufficiency, FCI generates a<\/span><\/p>\n<p><b>Partial Ancestral Graph (PAG)<\/b><span style=\"font-weight: 400;\">. A PAG represents the common features of all DAGs that are observationally equivalent given the observed variables, even if unmeasured confounders or selection bias are present.<\/span><span style=\"font-weight: 400;\">26<\/span><\/p>\n<p><span style=\"font-weight: 400;\">FCI also begins with an undirected graph and removes edges based on conditional independence tests. However, it employs more complex rules for identifying conditioning sets, such as using &#8220;possibly d-separating&#8221; sets, and for orienting edges, which explicitly account for the potential influence of unobserved variables.<\/span><span style=\"font-weight: 400;\">16<\/span><span style=\"font-weight: 400;\"> It can identify &#8220;maximal ancestral graphs&#8221; (MAGs), which represent independence relations in the presence of latent and selection variables.<\/span><span style=\"font-weight: 400;\">27<\/span><span style=\"font-weight: 400;\"> A notable feature of FCI is its &#8220;anytime&#8221; property: the algorithm can be interrupted at any stage, and the output will still be correct in the large sample limit, though potentially less informative.<\/span><span style=\"font-weight: 400;\">26<\/span><span style=\"font-weight: 400;\"> Despite its robustness to hidden variables, FCI can be computationally very expensive. The number of conditional independence tests performed grows exponentially with the number of variables in the worst case, impacting both speed and accuracy, especially with smaller sample sizes, as tests conditional on many variables have low statistical power.<\/span><span style=\"font-weight: 400;\">26<\/span><span style=\"font-weight: 400;\"> FCI is crucial for real-world applications where the strong causal sufficiency assumption of the PC algorithm is almost always violated. It provides a more realistic and robust framework for causal discovery by acknowledging the inherent limitations of observational data and providing a less definitive but more trustworthy output (PAG) that accounts for unobserved factors.<\/span><\/p>\n<p>&nbsp;<\/p>\n<h3><b>Score-Based Methods<\/b><\/h3>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400;\">Score-based algorithms conceptualize causal discovery as an optimization problem. They define a &#8220;score function&#8221; that quantifies how well a given causal graph fits the observed data, typically incorporating a penalty for model complexity to prevent overfitting.<\/span><span style=\"font-weight: 400;\">17<\/span><span style=\"font-weight: 400;\"> The algorithm then systematically searches the vast space of possible graphs to identify the one that maximizes this score. A key advantage of this approach is that it inherently avoids the multiple testing problem that is often encountered in constraint-based methods.<\/span><span style=\"font-weight: 400;\">29<\/span><\/p>\n<p>&nbsp;<\/p>\n<h4><b>Greedy Equivalence Search (GES)<\/b><\/h4>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400;\">Greedy Equivalence Search (GES) is a widely used heuristic score-based algorithm that navigates the space of causal Bayesian networks to find the model with the highest Bayesian score. It operates in two principal phases:<\/span><\/p>\n<ul>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><b>Forward Search<\/b><span style=\"font-weight: 400;\">: Beginning with an empty graph, GES iteratively and greedily adds edges between nodes. Each potential addition is evaluated based on whether it increases the Bayesian score of the model. This process continues until no single edge addition can further improve the Bayesian score.<\/span><span style=\"font-weight: 400;\">17<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><b>Backward Search<\/b><span style=\"font-weight: 400;\">: Following the forward phase, GES transitions to a backward search, during which edges are removed from the graph. Similar to the forward phase, an edge is removed only if its removal leads to an increase in the Bayesian score. This backward process persists until no single edge removal can further enhance the score.<\/span><span style=\"font-weight: 400;\">17<\/span><\/li>\n<\/ul>\n<p><span style=\"font-weight: 400;\">GES operates under several critical assumptions: <\/span><b>i.i.d. (independent and identically distributed) observational samples<\/b><span style=\"font-weight: 400;\">, <\/span><b>linear relationships between variables with Gaussian noise terms<\/b><span style=\"font-weight: 400;\">, adherence to the <\/span><b>Causal Markov condition<\/b><span style=\"font-weight: 400;\">, <\/span><b>Faithfulness<\/b><span style=\"font-weight: 400;\">, and, importantly, the assumption of <\/span><b>no hidden confounders<\/b><span style=\"font-weight: 400;\">.<\/span><span style=\"font-weight: 400;\">17<\/span><span style=\"font-weight: 400;\"> A notable limitation is that standard implementations of GES do not support multi-processing.<\/span><span style=\"font-weight: 400;\">17<\/span><span style=\"font-weight: 400;\"> Furthermore, edge orientation in tabular data is generally less reliable compared to time series data, where the inherent temporal order provides valuable additional causal information that can aid in directing edges.<\/span><span style=\"font-weight: 400;\">17<\/span><span style=\"font-weight: 400;\"> GES offers a compelling alternative to constraint-based methods, particularly when the data characteristics align with its underlying assumptions (e.g., linearity, absence of hidden confounders). Its two-phase search strategy efficiently explores the graph space to find a locally optimal structure, showcasing the utility of a score-based optimization approach.<\/span><\/p>\n<p>&nbsp;<\/p>\n<h4><b>NOTEARS (Non-linear Optimization for Causal Discovery)<\/b><\/h4>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400;\">NOTEARS (Non-linear Optimization for Causal Discovery) represents a significant innovation by transforming the traditionally discrete and combinatorial problem of structural learning into a continuous, differentiable numerical optimization problem.<\/span><span style=\"font-weight: 400;\">22<\/span><span style=\"font-weight: 400;\"> This transformation enables the use of powerful gradient-based optimization techniques to identify the globally optimal causal graph (DAG).<\/span><span style=\"font-weight: 400;\">22<\/span><span style=\"font-weight: 400;\"> A key innovation lies in its use of a smooth, differentiable acyclicity constraint function, specifically<\/span><\/p>\n<p><span style=\"font-weight: 400;\">h(W) = tr(exp(W \u2218 W)) &#8211; d = 0, which mathematically ensures that the learned graph is a Directed Acyclic Graph (DAG) by penalizing cycles.<\/span><span style=\"font-weight: 400;\">22<\/span><span style=\"font-weight: 400;\"> The optimization problem is typically solved using augmented Lagrangian methods, converting a constrained problem into an unconstrained one solvable by numerical methods.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">While the core NOTEARS framework is flexible, its basic applications often implicitly or explicitly assume <\/span><b>Acyclicity<\/b><span style=\"font-weight: 400;\"> (which it enforces), <\/span><b>smoothness\/differentiability<\/b><span style=\"font-weight: 400;\"> of the loss function and constraint for gradient-based optimization, and sometimes <\/span><b>linearity<\/b><span style=\"font-weight: 400;\"> and <\/span><b>Gaussian noise<\/b><span style=\"font-weight: 400;\"> for simplicity. <\/span><b>Causal sufficiency<\/b><span style=\"font-weight: 400;\"> is also typically assumed in its basic application.<\/span><span style=\"font-weight: 400;\">14<\/span><span style=\"font-weight: 400;\"> A critical limitation of NOTEARS is its<\/span><\/p>\n<p><b>lack of scale-invariance<\/b><span style=\"font-weight: 400;\">. Research has demonstrated that a simple rescaling of input variables can significantly alter the derived DAG, suggesting that its results may not reflect true causal relationships but rather depend on arbitrary data scaling.<\/span><span style=\"font-weight: 400;\">31<\/span><span style=\"font-weight: 400;\"> This raises serious concerns about its robustness and generalizability in practical applications. Despite this, NOTEARS has found application in<\/span><\/p>\n<p><b>Feature Selection (FSNT)<\/b><span style=\"font-weight: 400;\">, where it identifies direct causal relationships between features and a target variable and then ranks features by &#8220;causal strength&#8221; to select an optimal subset.<\/span><span style=\"font-weight: 400;\">22<\/span><span style=\"font-weight: 400;\"> NOTEARS pioneered a new class of causal discovery algorithms by framing the problem as a continuous optimization task. However, its scale-invariance issue underscores the importance of robust algorithmic properties that extend beyond mere mathematical tractability for reliable causal discovery.<\/span><\/p>\n<p>&nbsp;<\/p>\n<h2><b>IV. Conclusion<\/b><\/h2>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400;\">Causal graph learning from observational data is a field of immense importance, driven by the need to move beyond mere correlations to understand the fundamental cause-and-effect relationships that govern complex systems. This understanding is critical for informed decision-making, effective interventions, and scientific discovery across diverse domains, from healthcare and economics to social sciences.<\/span><span style=\"font-weight: 400;\">1<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The pursuit of causal knowledge from observational data is inherently challenging due to the &#8220;fundamental problem of causal inference&#8221;\u2014the impossibility of observing counterfactuals directly.<\/span><span style=\"font-weight: 400;\">3<\/span><span style=\"font-weight: 400;\"> This necessitates reliance on strong, often untestable, identification assumptions such as exchangeability, consistency, positivity, causal sufficiency, the causal Markov condition, faithfulness, and acyclicity.<\/span><span style=\"font-weight: 400;\">4<\/span><span style=\"font-weight: 400;\"> The pervasive nature of biases\u2014including confounding, selection bias (often a form of collider bias), and measurement bias\u2014further complicates this endeavor.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> These biases are not isolated phenomena but often intricately interconnected, where a single flawed methodological choice can cascade into multiple distortions. Directed Acyclic Graphs (DAGs) are indispensable tools for diagnosing this interplay of biases and guiding appropriate covariate selection, moving beyond simplistic &#8220;control&#8221; practices to ensure the validity of causal conclusions.<\/span><span style=\"font-weight: 400;\">3<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The inherent &#8220;fragility&#8221; of relying on strong assumptions, which are frequently violated in real-world observational settings, underscores the limitations of drawing definitive causal conclusions from single studies. This fragility necessitates the adoption of &#8220;triangulation of evidence,&#8221; integrating results from multiple, diverse approaches, each with different and largely unrelated sources of potential bias, to build a stronger and more robust body of causal evidence.<\/span><span style=\"font-weight: 400;\">2<\/span><span style=\"font-weight: 400;\"> This approach shifts the focus from seeking a single &#8220;true&#8221; causal graph to understanding the robustness and sensitivity of causal claims across various modeling assumptions and methods.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Algorithmic advancements in causal discovery offer diverse strategies to tackle these challenges. Constraint-based methods like the PC algorithm leverage conditional independence tests to infer graph structures, while the FCI algorithm extends this to handle hidden variables and selection bias, providing a more realistic framework for complex real-world data.<\/span><span style=\"font-weight: 400;\">14<\/span><span style=\"font-weight: 400;\"> Score-based methods, such as GES, frame causal discovery as an optimization problem, searching for graphs that best fit the data according to a defined score.<\/span><span style=\"font-weight: 400;\">17<\/span><span style=\"font-weight: 400;\"> Innovations like NOTEARS have transformed structural learning into a continuous optimization problem, though its lack of scale-invariance highlights the importance of robust algorithmic properties beyond mere mathematical tractability.<\/span><span style=\"font-weight: 400;\">22<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Despite significant progress, particularly with the integration of machine learning and the emerging role of Large Language Models (LLMs) in direct causal extraction from text, integrating domain knowledge, and refining causal structures <\/span><span style=\"font-weight: 400;\">38<\/span><span style=\"font-weight: 400;\">, several challenges persist. The difficulty in evaluating the quality of discovered causal structures in real-world datasets due to the absence of known ground truth remains a primary limitation, often necessitating reliance on synthetic data for evaluation.<\/span><span style=\"font-weight: 400;\">19<\/span><span style=\"font-weight: 400;\"> Furthermore, neural causal discovery methods still face accuracy issues in reliably distinguishing between existing and non-existing causal relationships in finite sample regimes, with the faithfulness property identified as a critical bottleneck.<\/span><span style=\"font-weight: 400;\">19<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Future directions in causal inference research are poised to address these limitations. Key areas include developing methods for high-dimensional data, particularly in confounding and mediation, where complex variable selection and interaction effects pose significant hurdles.<\/span><span style=\"font-weight: 400;\">37<\/span><span style=\"font-weight: 400;\"> There is a growing emphasis on precision medicine, causal machine learning (aiming to predict outcomes under interventions), enriching randomized experiments with real-world data, and addressing algorithmic fairness and social responsibility through a causal lens.<\/span><span style=\"font-weight: 400;\">37<\/span><span style=\"font-weight: 400;\"> Distributed learning, interference and spillover effects, and transportability of causal effects across populations are also active areas of research.<\/span><span style=\"font-weight: 400;\">37<\/span><span style=\"font-weight: 400;\"> The integration of explainability techniques into causal discovery, as seen in methods like ReX, represents a promising new direction for enhancing interpretability and bridging the gap between predictive modeling and causal inference.<\/span><span style=\"font-weight: 400;\">44<\/span><span style=\"font-weight: 400;\"> Ultimately, advancing causal graph learning in observational data will require continued innovation in algorithmic design, rigorous validation against realistic benchmarks, and a deep understanding of the underlying assumptions and their implications.<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Causal Graph Learning in Observational Data I. Introduction to Causal Graph Learning in Observational Data Causal graph learning represents a pivotal subfield within machine learning, shifting the focus from mere <span class=\"readmore\"><a href=\"https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/\">Read More &#8230;<\/a><\/span><\/p>\n","protected":false},"author":2,"featured_media":0,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[2019,1297,5],"tags":[],"class_list":["post-3104","post","type-post","status-publish","format-standard","hentry","category-big-data-2","category-data-engineer","category-infographics"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.4 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>Causal Graph Learning in Observational Data | Uplatz Blog<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Causal Graph Learning in Observational Data | Uplatz Blog\" \/>\n<meta property=\"og:description\" content=\"Causal Graph Learning in Observational Data I. Introduction to Causal Graph Learning in Observational Data Causal graph learning represents a pivotal subfield within machine learning, shifting the focus from mere Read More ...\" \/>\n<meta property=\"og:url\" content=\"https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/\" \/>\n<meta property=\"og:site_name\" content=\"Uplatz Blog\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/Uplatz-1077816825610769\/\" \/>\n<meta property=\"article:published_time\" content=\"2025-06-27T09:39:01+00:00\" \/>\n<meta name=\"author\" content=\"uplatzblog\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@uplatz_global\" \/>\n<meta name=\"twitter:site\" content=\"@uplatz_global\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"uplatzblog\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"22 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/causal-graph-learning-in-observational-data\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/causal-graph-learning-in-observational-data\\\/\"},\"author\":{\"name\":\"uplatzblog\",\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/#\\\/schema\\\/person\\\/8ecae69a21d0757bdb2f776e67d2645e\"},\"headline\":\"Causal Graph Learning in Observational Data\",\"datePublished\":\"2025-06-27T09:39:01+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/causal-graph-learning-in-observational-data\\\/\"},\"wordCount\":4774,\"publisher\":{\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/#organization\"},\"articleSection\":[\"Big Data\",\"Data Engineer\",\"Infographics\"],\"inLanguage\":\"en-US\"},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/causal-graph-learning-in-observational-data\\\/\",\"url\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/causal-graph-learning-in-observational-data\\\/\",\"name\":\"Causal Graph Learning in Observational Data | Uplatz Blog\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/#website\"},\"datePublished\":\"2025-06-27T09:39:01+00:00\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/causal-graph-learning-in-observational-data\\\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/uplatz.com\\\/blog\\\/causal-graph-learning-in-observational-data\\\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/causal-graph-learning-in-observational-data\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Causal Graph Learning in Observational Data\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/#website\",\"url\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/\",\"name\":\"Uplatz Blog\",\"description\":\"Uplatz is a global IT Training &amp; Consulting company\",\"publisher\":{\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/#organization\",\"name\":\"uplatz.com\",\"url\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/wp-content\\\/uploads\\\/2016\\\/11\\\/Uplatz-Logo-Copy-2.png\",\"contentUrl\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/wp-content\\\/uploads\\\/2016\\\/11\\\/Uplatz-Logo-Copy-2.png\",\"width\":1280,\"height\":800,\"caption\":\"uplatz.com\"},\"image\":{\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/#\\\/schema\\\/logo\\\/image\\\/\"},\"sameAs\":[\"https:\\\/\\\/www.facebook.com\\\/Uplatz-1077816825610769\\\/\",\"https:\\\/\\\/x.com\\\/uplatz_global\",\"https:\\\/\\\/www.instagram.com\\\/\",\"https:\\\/\\\/www.linkedin.com\\\/company\\\/7956715?trk=tyah&amp;amp;amp;amp;trkInfo=clickedVertical:company,clickedEntityId:7956715,idx:1-1-1,tarId:1464353969447,tas:uplatz\"]},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/uplatz.com\\\/blog\\\/#\\\/schema\\\/person\\\/8ecae69a21d0757bdb2f776e67d2645e\",\"name\":\"uplatzblog\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/7f814c72279199f59ded4418a8653ad15f5f8904ac75e025a4e2abe24d58fa5d?s=96&d=mm&r=g\",\"url\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/7f814c72279199f59ded4418a8653ad15f5f8904ac75e025a4e2abe24d58fa5d?s=96&d=mm&r=g\",\"contentUrl\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/7f814c72279199f59ded4418a8653ad15f5f8904ac75e025a4e2abe24d58fa5d?s=96&d=mm&r=g\",\"caption\":\"uplatzblog\"}}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Causal Graph Learning in Observational Data | Uplatz Blog","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/","og_locale":"en_US","og_type":"article","og_title":"Causal Graph Learning in Observational Data | Uplatz Blog","og_description":"Causal Graph Learning in Observational Data I. Introduction to Causal Graph Learning in Observational Data Causal graph learning represents a pivotal subfield within machine learning, shifting the focus from mere Read More ...","og_url":"https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/","og_site_name":"Uplatz Blog","article_publisher":"https:\/\/www.facebook.com\/Uplatz-1077816825610769\/","article_published_time":"2025-06-27T09:39:01+00:00","author":"uplatzblog","twitter_card":"summary_large_image","twitter_creator":"@uplatz_global","twitter_site":"@uplatz_global","twitter_misc":{"Written by":"uplatzblog","Est. reading time":"22 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/#article","isPartOf":{"@id":"https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/"},"author":{"name":"uplatzblog","@id":"https:\/\/uplatz.com\/blog\/#\/schema\/person\/8ecae69a21d0757bdb2f776e67d2645e"},"headline":"Causal Graph Learning in Observational Data","datePublished":"2025-06-27T09:39:01+00:00","mainEntityOfPage":{"@id":"https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/"},"wordCount":4774,"publisher":{"@id":"https:\/\/uplatz.com\/blog\/#organization"},"articleSection":["Big Data","Data Engineer","Infographics"],"inLanguage":"en-US"},{"@type":"WebPage","@id":"https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/","url":"https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/","name":"Causal Graph Learning in Observational Data | Uplatz Blog","isPartOf":{"@id":"https:\/\/uplatz.com\/blog\/#website"},"datePublished":"2025-06-27T09:39:01+00:00","breadcrumb":{"@id":"https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/uplatz.com\/blog\/causal-graph-learning-in-observational-data\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/uplatz.com\/blog\/"},{"@type":"ListItem","position":2,"name":"Causal Graph Learning in Observational Data"}]},{"@type":"WebSite","@id":"https:\/\/uplatz.com\/blog\/#website","url":"https:\/\/uplatz.com\/blog\/","name":"Uplatz Blog","description":"Uplatz is a global IT Training &amp; Consulting company","publisher":{"@id":"https:\/\/uplatz.com\/blog\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/uplatz.com\/blog\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/uplatz.com\/blog\/#organization","name":"uplatz.com","url":"https:\/\/uplatz.com\/blog\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/uplatz.com\/blog\/#\/schema\/logo\/image\/","url":"https:\/\/uplatz.com\/blog\/wp-content\/uploads\/2016\/11\/Uplatz-Logo-Copy-2.png","contentUrl":"https:\/\/uplatz.com\/blog\/wp-content\/uploads\/2016\/11\/Uplatz-Logo-Copy-2.png","width":1280,"height":800,"caption":"uplatz.com"},"image":{"@id":"https:\/\/uplatz.com\/blog\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/Uplatz-1077816825610769\/","https:\/\/x.com\/uplatz_global","https:\/\/www.instagram.com\/","https:\/\/www.linkedin.com\/company\/7956715?trk=tyah&amp;amp;amp;amp;trkInfo=clickedVertical:company,clickedEntityId:7956715,idx:1-1-1,tarId:1464353969447,tas:uplatz"]},{"@type":"Person","@id":"https:\/\/uplatz.com\/blog\/#\/schema\/person\/8ecae69a21d0757bdb2f776e67d2645e","name":"uplatzblog","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/secure.gravatar.com\/avatar\/7f814c72279199f59ded4418a8653ad15f5f8904ac75e025a4e2abe24d58fa5d?s=96&d=mm&r=g","url":"https:\/\/secure.gravatar.com\/avatar\/7f814c72279199f59ded4418a8653ad15f5f8904ac75e025a4e2abe24d58fa5d?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/7f814c72279199f59ded4418a8653ad15f5f8904ac75e025a4e2abe24d58fa5d?s=96&d=mm&r=g","caption":"uplatzblog"}}]}},"_links":{"self":[{"href":"https:\/\/uplatz.com\/blog\/wp-json\/wp\/v2\/posts\/3104","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/uplatz.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/uplatz.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/uplatz.com\/blog\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/uplatz.com\/blog\/wp-json\/wp\/v2\/comments?post=3104"}],"version-history":[{"count":2,"href":"https:\/\/uplatz.com\/blog\/wp-json\/wp\/v2\/posts\/3104\/revisions"}],"predecessor-version":[{"id":3127,"href":"https:\/\/uplatz.com\/blog\/wp-json\/wp\/v2\/posts\/3104\/revisions\/3127"}],"wp:attachment":[{"href":"https:\/\/uplatz.com\/blog\/wp-json\/wp\/v2\/media?parent=3104"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/uplatz.com\/blog\/wp-json\/wp\/v2\/categories?post=3104"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/uplatz.com\/blog\/wp-json\/wp\/v2\/tags?post=3104"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}