Intervention Design for Causal Representation Learning


In this paper, we take a first step towards bringing two fields of causality closer together, intervention design and causal representation learning. Intervention design is a well studied task in classic causal discovery, which aims at finding the minimal sets of experiments under which the causal graph can be identified. Causal representation learning aims at recovering causal variables from high-dimensional entangled observations. In recent work in causal representation, interventions are exploited to improve identifiability, similarly to classic causal discovery. Hence, the same task becomes relevant in this setting as well, how many experiments are minimally needed to identify the latent causal variables? Based on the recent causal representation learning method CITRIS, we show that for causal variables, experiments are sufficient to identify causal variables from temporal, intervened sequences, which is only one more experiment than needed for classic causal discovery in the worst case. Further, we show that this bound holds empirically in experiments on a 3D rendered video dataset.

UAI 2022 Workshop on Causal Representation Learning
Sara Magliacane
Sara Magliacane
Assistant Professor, Researcher

I’m an assistant professor at the University of Amsterdam and a researcher at the MIT-IBM Watson AI Lab. I work on causality, causal representation learning and causality-inspired ML.