The Fundamental Problem of Causal Inference
Consider the potential outcomes and , where denotes the outcome that unit (individual) would have if unit receives treatment . If you prefer do-notation, is the do-notation equivalent of . The fundamental problem of causal inference is that it is impossible to observe both and (Holland, 1986). This is because you cannot give unit treatment , observe , rewind time to before treatment was given, and give unit treatment to observe . We illustrate this below in a table with question marks in place of the potential outcomes that cannot be observed.
You might be thinking you can just give unit treatment to observe , wait some time, and then give unit treatment . However, the second potential outcome you observe will not necessarily be . Really, it is , which denotes the compound treatment of first giving treatment , then waiting, and then giving treatment . If you assume waiting sufficiently long nullifies the effects of the initial treatment , you get . If you further assume that all relevant confounding factors are exactly the same after waiting as they were when you gave the initial treatment , then you get . However, these are two strong assumptions.
This means that it is impossible to calculate the unit-level causal effect . This is why much of causal inference is focused on average treatment effects (ATE) , which are possible to calculate under assumptions such as SUTVA and ignorability (see, e.g., Morgan & Winship (2014, Sections 2.5 and 4.3.1)).
No Problem in Simulations
However, experiments in a virtual world such as a computer program don’t suffer from the fundamental problem of causal inference. Whenever our experiments are taking place in a program where we can change the treatment by changing the program, we can observe both and . To observe , we just run the program with
treatment = t. All we have to do to observe is modify the program by changing the assignment statement to
treatment = t' and rerun the program. It’s that simple. What are referred to as “simulations” in causal inference are just programs, so the fundamental problem of causal inference ain’t no problem there either.
This is not a particularly novel observation, but I think it is important to emphasize to machine learning researchers because so many of their experiments actually take place in computer simulations that they have a high degree of control over. I’ll present some highly simplified examples below. Think of these as seeds for you to come up with your own examples that could be much more complicated.
Supervised Learning Example
Consider a neural network where everything about the learning process is fixed except for the learning rate, so you are looking at a particular network (unit ). Your “treatment” is the learning rate . Your outcome of interest is the test error . You have already tried a learning rate of and observed . You are wondering what would happen if you were to change the learning rate to . Because you have control of the program, you just change the learning rate to , keeping everything else in the program the same. You then observe . This procedure may be very familiar to you. You bypassed the fundamental problem of causal inference and computed a unit-level causal effect: . The fact that this causal effect is negative is why you would choose a learning rate of instead of a learning rate of .
Reinforcement Learning Example
Consider a reinforcement learning agent in state at time (). I will omit the subscript , but it is implicit in this example. Think of the specific unit as specifying everything else that is held constant in the program such as the random seed, specific task, specific learner, etc. Say the action space is . These are your possible “treatments.” Your outcomes of interest are the reward and next state , which are both dependent on your action (potentially stochastically, depending on your environment). You can observe all 3 potential outcomes: . All you have to do is run the program 3 times from that particular state, taking a different action each time.
How About the Real World?
Of course, the fundamental problem of causal inference is still a problem outside of computer programs. However, the better we can model the world (in computers), the less of a problem the fundamental problem of causal inference becomes, in general. If we ever have good enough models of the world, we will be able to observe both and by simply rerunning the model with and with , just as I described above for a regular computer program. And modeling the world is an active research topic. For example, there was recently the NeurIPS 2018 Workshop on Modeling the Physical World: Learning, Perception, and Control.
If you care about causal effects in computer programs and you have access to the assignments for the causal variables you care about, you do not need to worry about the fundamental problem of causal inference. Because experiments in machine learning largely take place in computers, this had broad implications in this field.
Progress on modeling the physical world will directly translate to progress on estimating more and more accurate unit-level causal effects in the physical world.
- Holland, P. W. (1986). Statistics and Causal Inference. Journal of the American Statistical Association.
- Morgan, S. L., & Winship, C. (2014). Counterfactuals and Causal Inference: Methods and Principles for Social Research.