Elias Bareinboim
Wenzhuo Zhou
We study offline reinforcement learning (RL), which seeks to learn a good policy based on a fixed, pre-collected dataset. A fundamental challenge behind this task is the distributional shift due to the dataset lacking sufficient exploration, especially under function approximation. To tackle this issue, a bi-level structured policy optimization algorithm is proposed that models a hierarchical interaction between the policy (upper level) and the value function (lower level). The lower level focuses on constructing a confidence set of value estimates that maintain sufficiently small weighted average Bellman errors while controlling uncertainty arising from distribution mismatch. Subsequently, at the upper level, the policy aims to maximize a conservative value estimate from the confidence set formed at the lower level. This novel formulation preserves the maximum flexibility of the implicitly induced exploratory data distribution, enabling the power of model extrapolation. In practice, it can be solved through a computationally efficient, penalized adversarial estimation procedure. The theoretical regret guarantees do not rely on any data-coverage and completeness-type assumptions, only requiring realizability. These guarantees also demonstrate that the learned policy represents the best effort among all policies, as no other policies can outperform it.
YingYing Dong
This paper investigates distributional changes in the first stage to identify and estimate causal effects of a continuous endogenous variable (or treatment) using a binary or discrete instrument. We explore two alternative assumptions regarding the heterogeneity of the instrument's first-stage effect: LATE-type monotonicity and treatment rank similarity, which have been studied in separate strands of literature. Under treatment rank similarity, we derive simple estimands for average treatment effects at different treatment quantiles, capturing treatment effect heterogeneity across treatment levels. Additionally, we propose a doubly robust causal estimand that identifies a weighted average treatment effect for all units responsive to the instrument when either of these two non-nested assumptions holds. Our doubly robust framework subsumes LATE-type estimands as a special case. We also provide asymptotically normal semiparametric estimators based on these identification results and demonstrate the proposed methods in an empirical application estimating the effects of sleep on well-being.
Babak Salimi
ML models in critical applications like healthcare and finance often rely on incomplete, noisy, or biased data, where ground truth is unrecoverable. Conventional data cleaning falls short, leading to models that appear fair and accurate during development but fail in deployment. We first discuss an approach for learning fair models by leveraging background knowledge from the data collection process and external knowledge about the domain to enable learning fair models that are robust to incomplete data.
We then present a novel framework that systematically propagates uncertainty throughout the training pipeline to enhance model robustness when data is incomplete or uncertain. Our approach integrates Possible Worlds Semantics from database theory and utilizes abstract interpretation techniques to analyze robustness by propagating variations caused by incomplete and uncertain data into models and their decisions. This framework ensures a principled way to quantify and address uncertainty, enabling reliable predictions under varying data conditions.
Biwei Huang
Abstract: Recently, causality has garnered significant interest within the research communities of statistics, machine learning, and computer science. A central challenge in this field is uncovering the underlying causal structures and models. Traditional methods for causal structure learning often assume the absence of latent confounders. In this talk, I will highlight recent advances in causal structure learning that specifically address the challenges posed by latent confounders. I will focus on three key techniques and their associated structural or distributional constraints, which enable us to identify latent variables, determine their cardinalities, and map out the structure involving both latent and observed variables.
David Danks
Much of the research on causal discovery methods focuses on methods able to handle increasingly complex systems, whether in terms of number of variables, density of connections, complexity of parametric functions, and more. In contrast, this talk will focus on causal discovery when we have challenges due to complicated measurements. Specifically, I focus on two different types of challenges. First, we might not be able to measure at the same timescale as the causal connections; for example, fMRI measurements are much slower than the brain’s timescale, or economic measures are typically released less frequently than economic activity occurs. I will describe several causal discovery algorithms that we have developed for this situation, demonstrating both theoretical and real-world performance. I also demonstrate that we can sometimes benefit by measuring even more slowly, contra the standard advice that “faster measurement is better measurement.” Second, we might not be able to measure all of the relevant variables at the same time; for example, a healthcare organization and financial institution might each measure variables that interest them, but we cannot (for privacy reasons) integrate those into a single dataset. I describe causal discovery methods that can extract information about the global causal structure from these partial, overlapping datasets, and show that the methods can yield novel scientific insights on real-world data.
Judea Pearl
Hengrui Cai
The causal revolution has spurred interest in understanding complex relationships across various fields. Most existing methods aim to discover causal relationships among variables within a complex, large-scale system. However, in practice, only a small number of variables are relevant to the outcomes of interest. As a result, causal estimation using the full causal representation, especially with limited data, could lead to many falsely discovered, spurious variables that are highly correlated with but have no causal impact on the target outcome. We propose learning a class of necessary and sufficient causal representation that only contain causally relevant variables, utilizing probabilities of causation. The framework is further extended to natural language processing models to disentangle the 'black box' by identifying true rationales when two or more snippets are highly inter-correlated, thus contributing similarly to prediction accuracy. We leverage two causal desiderata, non-spuriousness and efficiency, establishing their theoretical identification as the main component of learning necessary and sufficient in language models. The superior performance of our proposed methods is demonstrated in real-world reviews and medical datasets through extensive experiments.
Mimi Liljeholm
Causal invariance – the extent to which a putative cause acts “the same” across the many constellations in which it occurs – is central to the generalization of human knowledge and the reproducibility of scientific results. In this talk, I will present work aimed at characterizing how human reasoners tacitly define invariant causal influence. Specifically, I will contrast an Associationist account, according to which invariance is defined strictly on observed occurrences of a putative cause and its effect, with a Causal Power account that defines invariance on latent causal influences. I will instantiate these competing definitions of invariance in a Bayesian Reinforcement Learning framework and assess how well they predict uncertainty in human causal induction, at behavioral and neural levels. I will argue that the results are consistent with a Causal Power account of causal reasoning.
Adrian Lozano-Duran
Causality lies at the heart of scientific inquiry, serving as the fundamental basis for understanding the interactions among variables in physical systems. Despite its central role, current methods for causal inference face significant challenges due to the presence of nonlinear dependencies, stochastic and deterministic interactions, self-causation, mediator, confounder, and collider effects, and contamination from unobserved, exogenous factors, to name a few. While there are methods that can effectively address some of these challenges, no single approach has been successful in integrating all these aspects. Here, we tackle these challenges with SURD: Synergistic-Unique-Redundant Decomposition of causality. SURD quantifies causality as the increments of redundant, unique, and synergistic information gained about future events based on available information from past observations. The formulation is non-intrusive and requires only pairs of past and future events, facilitating its application in both computational and experimental investigations, even when samples are scarce. We benchmark SURD against existing methods in scenarios that pose significant challenges in causal inference. These include synchronization in logistic maps, the Rössler-Lorenz system, the Lotka-Volterra prey-predator model, the Moran effect model, and energy cascade in turbulence, among others. Our findings demonstrate that SURD offers a more reliable quantification of causality compared to state-of-the-art methods for causal inference.
Tianchen Qian
To optimize mobile health interventions and advance domain knowledge on intervention design, it is critical to understand how the intervention effect varies over time and with contextual information. This study aims to assess how a push notification suggesting physical activity influences individuals’ step counts using data from the HeartSteps micro-randomized trial (MRT). The statistical challenges include the time-varying treatments and longitudinal functional step count measurements. We propose the first semiparametric causal excursion effect model with varying coefficients to model the time-varying effects within a decision point and across decision points in an MRT. The proposed model incorporates double time indices to accommodate the longitudinal functional outcome, enabling the assessment of time-varying effect moderation by contextual variables. We propose a two-stage causal effect estimator that is robust against a misspecified high-dimensional outcome regression nuisance model. We establish asymptotic theory and conduct simulation studies to validate the proposed estimator. Our analysis provides new insights into individuals’ change in response profiles (such as how soon a response occurs) due to the activity suggestions, how such changes differ by the type of suggestions received, and how such changes depend on other contextual information such as being recently sedentary and the day being a weekday.