COMING SOON
COMING SOON
COMING SOON
In the previous module on the potential outcomes model, we defined confounders as variables that are associated with both the treatment and the outcome. We called these criteria the common-cause criteria. We said that without considering them in causal analysis, the estimates of the causal effect might be biased. This bias is usually referred to as the confounding bias
.
Now that we are familiar with DAGs we can check the common-cause criteria in a visual way. Some examples may help. In all the examples below, we’d like to check whether variable is a confounder. In all these DAGs, represents the treatment and represents the outcome.
In our first DAG, let’s see if is a confounder!
Therefore, is not a confounder. Now, consider a second DAG shown below:
As a result, is a confounder based on the conditions mentioned above. Here’s another example.
Thus, is a confounder. Ok! One more example and we’ll be done! 😊
Again, is a confounder.
As it turns out, the common-cause criteria alone are not enough for identifying confounders. In fact, this traditional approach to causal inference (what we have followed so far) can lead to inappropriate adjustments.
It turns out we can use DAGs in a more appropriate way to identify confounding bias. To understand this, we need to learn about frontdoor and backdoor paths. We will see that to sufficiently control for confounding, we must block backdoor paths from the treatment to the outcome. If we block the backdoor path, then the ignorability assumption is satisfied. So instead of going after all confounders, we should only focus on confounders that create what are called backdoor paths
.
Let’s first start with a frontdoor path. In a causal diagram, a frontdoor path
is a path from treatment to the outcome that begins with an arrow out of the treatment and ends in the outcome. In the DAG below, if is the treatment and is the outcome, is a frontdoor path from the treatment to the outcome.
Frontdoor paths should not be blocked because they are what we’re interested in, to begin with. If a frontdoor path is blocked, then we won’t be able to see the direct effect of the treatment on the outcome.
Note that the common-cause criteria for finding confounders would tell us to control for because it’s associated with both the treatment and the outcome. However, if we control for , we block the path that would give us the treatment effect itself. In the example above, if we block the frontdoor path by controlling for , we won’t be able to find the effect of on . This is because partially captures the treatment.
Ok. Now that we know about frontdoors, let’s figure out what backdoor paths are. Backdoor paths
are paths between the treatment and the outcome but instead of going out of the treatment, the path is directed into the treatment. In the DAG above, is the only backdoor path. Note that the path is between the treatment and the outcome and it includes an arrow going towards the treatment.
While frontdoor paths capture the effect of the treatment on the outcome, backdoor paths represent alternative paths (channels) between the treatment and the outcome that have nothing to do with the treatment effect.
In causal inference, we like to estimate the causal effect of the treatment on the outcome and not any (conditional) associations caused by other variables. Take in the quiz above! If we don’t control for , some of the associations between and will be through the backdoor paths is on. So to estimate the pure causal effect of on , we need to block these backdoor paths.
Imagine we are interested in the effect of family socioeconomic status (SES) on individuals’ income. Therefore, SES is our treatment, and income is our outcome. For the sake of our example, imagine the following DAG represents the causal relationships.
In this DAG, we need to make sure we don’t block the frontdoor path (the causal path) and we block all backdoor paths. The interesting thing about this DAG is that there are no backdoor paths. So there is no need to block any backdoor paths. But try to identify the frontdoor path.
You’ll notice that there are four!
If we control for both education and job, we will likely close all of these frontdoor paths, and therefore, we’ll be unable to establish the causal effect of SES on incomes.
We will go over what should and shouldn’t be controlled in a DAG like this soon. The main lesson here is that the common-cause criteria aren’t the best methods for identifying what we should control for. A better approach is to identify backdoor paths and to block them.