PulseAugur
LIVE 03:33:53
research · [33 sources] ·
0
research

New AI methods advance causal discovery for complex, noisy, and large-scale data

Several recent arXiv papers introduce novel methods and benchmarks for causal discovery, a field focused on identifying cause-and-effect relationships from data. These advancements include techniques for handling noisy or incomplete data, integrating expert knowledge, and improving scalability for large datasets. New benchmarks and testing frameworks are also being developed to rigorously evaluate the robustness of existing causal discovery algorithms against various assumption violations, particularly in time-series data and natural language reasoning. AI

Summary written by gemini-2.5-flash-lite from 33 sources. How we write summaries →

IMPACT Advances in causal discovery methods could lead to more reliable AI systems capable of understanding and reasoning about cause-and-effect relationships, particularly in complex or noisy environments.

RANK_REASON Multiple arXiv papers published on May 7, 2026, detailing new methods and benchmarks for causal discovery.

Read on arXiv cs.CL →

COVERAGE [33]

  1. arXiv cs.LG TIER_1 · Francesco Locatello ·

    Causal Learning with the Invariance Principle

    Causal discovery, the problem of inferring the direction of causality, is generally ill-posed. We use the language of structural causal models (SCM) to show that assuming that the causal relations are acyclic and invariant across multiple environments (e.g., the way minimum wage …

  2. arXiv cs.AI TIER_1 · Francesco Locatello ·

    Towards a holistic understanding of Selection Bias for Causal Effect Identification

    Selection bias is pervasive in observational studies. For example, large scale biobanks data can exhibit ``healthy volunteer bias'' when respondents are healthier and of higher socio-economic status than the population they are meant to represent. Recovering causal effects from s…

  3. Hugging Face Daily Papers TIER_1 ·

    A Recursive Decomposition Framework for Causal Structure Learning in the Presence of Latent Variables

    Constraint-based causal discovery is widely used for learning causal structures, but heavy reliance on conditional independence (CI) testing makes it computationally expensive in high-dimensional settings. To mitigate this limitation, many divide-and-conquer frameworks have been …

  4. Hugging Face Daily Papers TIER_1 ·

    ConfoundingSHAP: Quantifying confounding strength in causal inference

    In causal inference, confounders are variables that influence both treatment decisions and outcomes. However, unlike as in randomized clinical trials, the treatment assignment mechanism in observational studies is not known, and it is thus unclear which covariates act as confound…

  5. arXiv cs.AI TIER_1 · Alex Markham ·

    Coarsening Linear Non-Gaussian Causal Models with Cycles

    Recent work on causal abstraction, in particular graphical approaches focusing on causal structure between clusters of variables, aims to summarize a high-dimensional causal structure in terms of a low-dimensional one. Existing methods for learning such summaries from data assume…

  6. arXiv cs.LG TIER_1 · Chris J. Maddison ·

    Bayesian Sensitivity of Causal Inference Estimators under Evidence-Based Priors

    Causal inference, especially in observational studies, relies on untestable assumptions about the true data-generating process. Sensitivity analysis helps us determine how robust our conclusions are when we alter these underlying assumptions. Existing frameworks for sensitivity a…

  7. arXiv cs.LG TIER_1 · Joseph D. Ramsey ·

    Fourier Feature Methods for Nonlinear Causal Discovery: FFML Scoring and FFCI Testing in Mixed Data

    arXiv:2605.05743v1 Announce Type: cross Abstract: Gaussian process marginal likelihood scores and kernel conditional independence tests are theoretically appealing for nonlinear causal discovery but computationally prohibitive at scale. We present two complementary RFF-based meth…

  8. arXiv cs.LG TIER_1 · Shicheng Fan, Nour Elhendawy, Jianle Sun, Ke Fang, Kun Zhang, Yihang Wang, Lu Cheng ·

    MOSAIC: Module Discovery via Sparse Additive Identifiable Causal Learning for Scientific Time Series

    arXiv:2605.05524v1 Announce Type: new Abstract: Causal representation learning (CRL) seeks to recover latent variables with identifiability guarantees, typically up to permutation and component-wise reparameterization under appropriate assumptions. However, identifiability does n…

  9. arXiv cs.LG TIER_1 · Sunmin Oh, Sang-Yun Oh, Gunwoong Park ·

    Relaxed Sparsest-Permutation Formulation for Causal Discovery at Scale

    arXiv:2605.05568v1 Announce Type: cross Abstract: Despite the growing availability of large datasets, causal structure learning remains computationally prohibitive at scale. We revisit sparsest-permutation learning for linear structural equation models and show that exact Cholesk…

  10. arXiv cs.LG TIER_1 · Adrick Tench, Thomas Demeester ·

    Dynamic Expert-Guided Model Averaging for Causal Discovery

    arXiv:2601.16715v2 Announce Type: replace Abstract: Would-be practitioners of causal discovery face a dizzying array of algorithms without a clear best choice. This abundance of competitive methods makes ensembling a natural strategy for practical applications. At the same time, …

  11. arXiv cs.LG TIER_1 · Marvin Sextro, Weronika K{\l}os, Gabriel Dernbach ·

    MapPFN: Learning Causal Perturbation Maps in Context

    arXiv:2601.21092v3 Announce Type: replace Abstract: Planning effective interventions in biological systems requires treatment-effect models that adapt to unseen biological contexts by identifying their specific underlying mechanisms. Yet single-cell perturbation datasets span onl…

  12. arXiv cs.LG TIER_1 · Thomas S. Robinson, Ranjit Lall ·

    PAIR-CI: Calibrated Conditional Independence Testing for Causal Discovery with Incomplete Data

    arXiv:2605.04838v1 Announce Type: cross Abstract: The standard constraint-based paradigm for causal discovery with incomplete data -- impute first, test second -- is frequently miscalibrated: any consistent conditional independence (CI) test rejects a true null with probability a…

  13. arXiv cs.LG TIER_1 · Geert Mesters, Alvaro Ribot, Anna Seigal, Piotr Zwiernik ·

    Causal discovery under mean independence and linearity

    arXiv:2605.04381v1 Announce Type: cross Abstract: Causal discovery methods such as LiNGAM identify causal structure from observational data by assuming mutually independent disturbances. This assumption is fragile: shared volatility, common scale effects, or other forms of depend…

  14. arXiv cs.CL TIER_1 · Zhi Xu, Yun Fu ·

    NoisyCausal: A Benchmark for Evaluating Causal Reasoning Under Structured Noise

    arXiv:2605.04313v1 Announce Type: new Abstract: Causal reasoning in natural language requires identifying relevant variables, understanding their interactions, and reasoning about effects and interventions, often under noisy or ambiguous conditions. While large language models (L…

  15. arXiv cs.LG TIER_1 · Bruno Petrungaro, Anthony C. Constantinou ·

    Time series causal discovery with variable lags

    arXiv:2605.04081v1 Announce Type: new Abstract: Causal Bayesian Networks (CBNs) are a powerful tool for reasoning under uncertainty about complex real-world problems. Such problems evolve over time, responding to external shocks as they occur. To support decision-making, CBNs req…

  16. arXiv cs.LG TIER_1 · Gideon Stein, Niklas Penzel, Tristan Piater, Joachim Denzler ·

    TCD-Arena: Assessing Robustness of Time Series Causal Discovery Methods Against Assumption Violations

    arXiv:2605.03045v1 Announce Type: new Abstract: Causal Discovery (CD) is a powerful framework for scientific inquiry. Yet, its practical adoption is hindered by a reliance on strong, often unverifiable assumptions and a lack of robust performance assessment. To address these limi…

  17. Hugging Face Daily Papers TIER_1 ·

    Causal discovery under mean independence and linearity

    Causal discovery methods such as LiNGAM identify causal structure from observational data by assuming mutually independent disturbances. This assumption is fragile: shared volatility, common scale effects, or other forms of dependence can cause the methods to recover the wrong ca…

  18. arXiv cs.CL TIER_1 · Yun Fu ·

    NoisyCausal: A Benchmark for Evaluating Causal Reasoning Under Structured Noise

    Causal reasoning in natural language requires identifying relevant variables, understanding their interactions, and reasoning about effects and interventions, often under noisy or ambiguous conditions. While large language models (LLMs) exhibit strong general reasoning abilities,…

  19. arXiv stat.ML TIER_1 · Jin Du, Li Chen, Xun Xian, An Luo, Fangqiao Tian, Ganghua Wang, Charles Doss, Xiaotong Shen, Jie Ding ·

    Ice Cream Doesn't Cause Drowning: Benchmarking LLMs Against Statistical Pitfalls in Causal Inference

    arXiv:2505.13770v3 Announce Type: replace-cross Abstract: Reliable causal inference is essential for making decisions in high-stakes areas like medicine, economics, and public policy. However, it remains unclear whether large language models (LLMs) can handle rigorous and trustwo…

  20. arXiv stat.ML TIER_1 · Oliver J. Hines, Caleb H. Miles ·

    Learning density ratios in causal inference using Bregman-Riesz regression

    arXiv:2510.16127v2 Announce Type: replace Abstract: The ratio of two probability density functions is a fundamental quantity that appears in many areas of statistics and machine learning, including causal inference, reinforcement learning, covariate shift, outlier detection, inde…

  21. arXiv stat.ML TIER_1 · Hao Zhang ·

    A Recursive Decomposition Framework for Causal Structure Learning in the Presence of Latent Variables

    Constraint-based causal discovery is widely used for learning causal structures, but heavy reliance on conditional independence (CI) testing makes it computationally expensive in high-dimensional settings. To mitigate this limitation, many divide-and-conquer frameworks have been …

  22. arXiv stat.ML TIER_1 · Stefan Feuerriegel ·

    Amortizing Causal Sensitivity Analysis via Prior Data-Fitted Networks

    Causal sensitivity analysis aims to provide bounds for causal effect estimates in the presence of unobserved confounding. However, existing methods for causal sensitivity analysis are per-instance procedures, meaning that changes to the dataset, causal query, sensitivity level, o…

  23. arXiv stat.ML TIER_1 (CA) · Tobias Maringgele, Jalal Etesami ·

    Optimal Experiments for Partial Causal Effect Identification

    arXiv:2605.06993v1 Announce Type: cross Abstract: Causal queries are often only partially identifiable from observational data, and experiments that could tighten the resulting bounds are typically costly. We study the problem of selecting, prior to observing experimental outcome…

  24. arXiv stat.ML TIER_1 · Shakeel Gavioli-Akilagun, Kieran Wood, Francesco Quinzan ·

    Detecting Changes in Causal Dependence with Kernels and Copulas

    arXiv:2605.05809v1 Announce Type: cross Abstract: We propose a framework for determining whether the causal dependence of an outcome $Y$ on a covariate $X$ changes at a given time point, given confounders $\boldsymbol{Z}$. For instance, in financial markets, the effect of a marke…

  25. arXiv stat.ML TIER_1 (CA) · Jalal Etesami ·

    Optimal Experiments for Partial Causal Effect Identification

    Causal queries are often only partially identifiable from observational data, and experiments that could tighten the resulting bounds are typically costly. We study the problem of selecting, prior to observing experimental outcomes, a cost-constrained subset of experiments that m…

  26. arXiv stat.ML TIER_1 · Francesco Quinzan ·

    Detecting Changes in Causal Dependence with Kernels and Copulas

    We propose a framework for determining whether the causal dependence of an outcome $Y$ on a covariate $X$ changes at a given time point, given confounders $\boldsymbol{Z}$. For instance, in financial markets, the effect of a market indicator on asset returns may causally change o…

  27. arXiv stat.ML TIER_1 · Joseph D. Ramsey ·

    Fourier Feature Methods for Nonlinear Causal Discovery: FFML Scoring and FFCI Testing in Mixed Data

    Gaussian process marginal likelihood scores and kernel conditional independence tests are theoretically appealing for nonlinear causal discovery but computationally prohibitive at scale. We present two complementary RFF-based methods forming a practical toolkit for score-based, c…

  28. arXiv stat.ML TIER_1 · Gunwoong Park ·

    Relaxed Sparsest-Permutation Formulation for Causal Discovery at Scale

    Despite the growing availability of large datasets, causal structure learning remains computationally prohibitive at scale. We revisit sparsest-permutation learning for linear structural equation models and show that exact Cholesky factorization is unnecessary for structure recov…

  29. arXiv stat.ML TIER_1 · Ranjit Lall ·

    PAIR-CI: Calibrated Conditional Independence Testing for Causal Discovery with Incomplete Data

    The standard constraint-based paradigm for causal discovery with incomplete data -- impute first, test second -- is frequently miscalibrated: any consistent conditional independence (CI) test rejects a true null with probability approaching 1 when imputation error induces spuriou…

  30. arXiv stat.ML TIER_1 · Piotr Zwiernik ·

    Causal discovery under mean independence and linearity

    Causal discovery methods such as LiNGAM identify causal structure from observational data by assuming mutually independent disturbances. This assumption is fragile: shared volatility, common scale effects, or other forms of dependence can cause the methods to recover the wrong ca…

  31. arXiv stat.ML TIER_1 · Xihang Shan, Da Zhou ·

    PRCD-MAP: Learning How Much to Trust Imperfect Priors in Causal Discovery

    arXiv:2605.01669v1 Announce Type: new Abstract: External priors of unknown reliability create a brittle trade-off in causal discovery: blind trust amplifies errors, blind rejection wastes signal. Real priors are also \emph{heterogeneously} reliable -- physical laws are trustworth…

  32. arXiv stat.ML TIER_1 · Da Zhou ·

    PRCD-MAP: Learning How Much to Trust Imperfect Priors in Causal Discovery

    External priors of unknown reliability create a brittle trade-off in causal discovery: blind trust amplifies errors, blind rejection wastes signal. Real priors are also \emph{heterogeneously} reliable -- physical laws are trustworthy, LLM-suggested edges are speculative -- yet ex…

  33. Towards AI TIER_1 · Ruiz Rivera ·

    Rethinking Predictors: Why Causal Reasoning Matters in Data Science (Part 1)

    <figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*nNbj5vzm6X-yFgapvo2g5g.jpeg" /><figcaption>Photo by <a href="https://unsplash.com/@theshubhamdhage?utm_source=unsplash&amp;utm_medium=referral&amp;utm_content=creditCopyText">Shubham Dhage</a> on <a href="https:/…