Transparent and Robust Causal Inference in Data Science

Event Start
Event End
Location
https://kaust.zoom.us/j/96131500026
Carlos Cinelli, Ph.D. candidate, Department of Statistics, UCLA

Abstract

The past few decades have witnessed rapid and unprecedented theoretical progress on the science of causal inference, ranging from the “credibility revolution” with the popularization of quasi-experimental designs, to the development of a complete solution to non-parametric identification with causal graphical models. Most of these theoretical progress, however, relies on strong, exact assumptions, such as the absence of unobserved common causes, or the absence of certain direct effects. Unfortunately, more often than not these assumptions are very hard to defend in practice. This leads to two undesirable consequences for applied quantitative work in the data-intensive sciences: (i) important research questions may be neglected, simply because they do not exactly match the requirements of current methods; or, (ii) researchers may succumb to making the required “identification assumptions” simply to justify the use of available methods, but not because these assumptions are truly believed (or understood).  In this talk, I will discuss new theories, methods, and software for permitting causal inferences under more flexible and realistic settings. These tools empower scientists, and policymakers to both examine the sensitivity of causal inferences to violations of its underlying assumptions, and also to draw robust and trustworthy conclusions from settings in which traditional methods fail.  

Brief Biography

Carlos Cinelli is a Ph.D. candidate in Statistics at the University of California, Los Angeles, working under the supervision of Judea Pearl and Chad Hazlett. His research focuses on developing new causal and statistical methods for transparent and robust causal claims in the empirical sciences. He is also interested in the intersections of causality with machine learning and artificial intelligence.

Contact Person