Mediators are variables that transmit causal effects from treatments to outcomes. Those who undertake mediation analysis seek
to answer “how” questions about causation: how does this treatment affect that outcome? Typically, we desire answers of the
form “the treatment affects a causally intermediate variable, which in turn affects the outcome.” Identifying these causally
intermediate variables is the challenge of mediation analysis.
We begin by characterizing the role that mediation analysis plays in political science. We then describe conventional methods
of mediation analysis and the bias to which they are prone. We proceed by describing experimental methods that can reliably
produce accurate estimates of mediation effects. The experimental approach has several important limitations, and we end the
section by explaining how these limitations imply both best practices and an agenda for future research. We consider objections
to our argument in the next section, including the common objection that manipulation of mediators is often infeasible. Our
last section reviews and concludes.
Like many related procedures, the method proposed by Baron and Kenny (
1986) is based on three models:
where
Y is the outcome of interest,
X is a treatment,
M is a potential mediator of
the treatment, and
α1,
α2, and
α3 are intercepts. For simplicity, we assume that
X and
M are binary variables coded either 0 or 1. The unobservable disturbances
e1,
e2, and
e3 are mean-zero error terms that represent the cumulative effect of omitted variables. It is not difficult to extend this framework
to include multiple mediators and other covariates, and our criticisms apply with equal force to models that include such
variables. For notational clarity and comparability to previous articles about mediation analysis, we limit our discussion
to the three-variable regression framework.
For simplicity, we assume throughout this chapter that X is randomly assigned such that it is independent of the disturbances: e1, e2, e3 ⊥⊥X. As we shall see, randomization of X alone does not ensure unbiased estimation of the effects of mediators. Consequently, we refer to designs in which only X is randomized as nonexperimental for the purpose of mediation analysis, reserving experimental for studies in which both X and M are randomized.
The coefficients of interest are
a,
b,
c, and
d. The total effect of
X on
Y is
c. To see how
c is typically decomposed into “direct” and “indirect” effects, substitute
Equation (1) into
Equation (3), yielding
The direct effect of
X is
d. The indirect or “mediated” effect is
ab (or, equivalently,
c − d).
2
Baron and Kenny (
1986) do not say how the coefficients in these equations are to be estimated; in practice, ordinary least squares (OLS) is almost
universally used. But the OLS estimator of
b in
Equation (3) is biased:
The OLS estimator of
d is also biased:
(A proof is given in
Bullock, Green, and Ha [
2008, 39–40].) OLS estimators of direct and indirect effects will therefore be biased as well.
In expectation, the OLS estimators of
b and
d produce accurate estimates only if cov(
e1,
e3) = 0.
3 But this condition is unlikely to hold unless both
X and
M are randomly assigned. The problem is straightforward: if an unobserved variable affects both
M and
Y, it will cause
e1 and
e3 to covary. And even if no unobserved variable affects both
M and
Y, these disturbances are likely to covary if
M is merely correlated with an unobserved variable that affects
Y, e.g., another mediator. This “multiple-mediator problem” is a serious threat to social-science mediation analysis because
most of the effects that interest social scientists are likely to have multiple correlated mediators. Indeed, we find it difficult
to think of any political effects that do not fit this description.
4
The standard temptation in nonexperimental analysis is to combat this problem by controlling for potential mediators other
than
M. But it is normally impossible to measure all possible mediators. Indeed, it may be impossible to merely
think of all possible mediators. And controlling for some potential mediators but not all of them is no guarantee of better estimates;
to the contrary, it may make estimates worse
(Clarke
2009). Fighting
endogeneity in nonexperimental mediation analysis by adding control variables is a method with no clear stopping rule or way
to detect bias – a shaky foundation on which to build beliefs about mediation.
Political scientists who use the Baron-Kenny (
1986) method and related methods often want to test hypotheses about several potential mediators rather than one. In these cases,
the most common approach is “one-at-a-time” estimation, whereby
Equation (3) is estimated separately for each mediator. This practice makes biased inferences about mediation even more likely. The researcher,
who already faces the spectre of bias due to the omission of variables over which she has no control, compounds the problem
by intentionally omitting variables that are likely to be important confounds. Nonexperimental mediation analysis is problematic
enough, but one-at-a-time testing of mediators stands out as an especially bad practice.
The Baron-Kenny method and related methods are often applied to experiments in which the treatment has been randomized but
the mediator has not, and there seems to be a widespread belief that such experiments are sufficient to ensure unbiased estimates
of direct and indirect effects. But randomization of the treatment is not enough to protect researchers from biased estimates.
It can ensure that
X bears no systematic relationship to
e1,
e2, or
e3, but it says nothing about whether
M is systematically related to those variables, and thus nothing about whether cov(
e1,
e3) = 0.
5
Stepping back from mediation analysis to the more general problem of estimating causal effects, note that estimators tend
to be biased when one controls for variables that are affected by the treatment. One does this whenever one controls for
M in a regression of
Y on
X, which the Baron-Kenny method requires. This “post-treatment bias” has been discussed in statistics and political science
(e.g., Rosenbaum
1984, 188–94;
King and Zeng
2006, 146–48), but its relevance to mediation analysis has gone largely unnoticed. At root, it is one instance of an even more
general rule: estimators of the parameters of regression equations are likely to be unbiased only if the predictors in those
equations are independent of the disturbances. And in most cases, the only way to ensure that
M is independent of the disturbances is to randomly assign its values. By contrast, “the benefits of randomization are generally
destroyed by including post-treatment variables”
(Gelman and Hill
2007, 192).
Within the past decade, statisticians and political scientists have advanced several different methods of mediation analysis
that do not call for manipulation of mediators. These methods improve on
Baron and Kenny (
1986), but they do not overcome the problem of endogeneity in nonexperimental mediation analysis. For example,
Frangakis and Rubin (
2002) propose “principal stratification,” which entails dividing subjects into groups on the basis of their potential outcomes
for mediators. Causal effects are then estimated separately for each “principal stratum.” The problem is that some potential
outcomes for each subject are necessarily unobserved, and those who use principal stratification must infer the values of
these potential outcomes on the basis of covariates. In practice, “this reduces to making the same kinds of assumptions as
are made in typical observational studies when ignorability is assumed” (Gelman and Hill
2007, 193).
In a different vein,
Imai, Keele, and Yamamoto (
2010) show that indirect effects can be identified even when the mediator is not randomized – provided that we stipulate the size
of cov(
e1,
e3). This is helpful: if we are willing to make assumptions about the covariance of unobservables, then we may be able to place
bounds on the likely size of the indirect effect. But in no sense is this method a substitute for experimental manipulation
of the mediator. Instead, it requires us to make strong assumptions about the properties of unobservable disturbances, just
as other methods do when they are applied to nonexperimental data. Moreover,
Imai, Keele, Tingley, and Yamamoto (
2010, 43) note that even if we are willing to stipulate the value of cov(
e1,
e3), the method that they propose cannot be used whenever the mediator of interest is directly affected by both the treatment
and another mediator. This point is crucial because many effects that interest political scientists seem likely to be transmitted
by multiple mediators that affect each other.
None of these warnings implies that all nonexperimental mediation research is equally suspect. All else equal, research in
which only a treatment is randomized is preferable to research in which no variables are randomized; treatment-only randomization
does not make accurate mediation inference likely, but it does clarify the assumptions required for accurate inference. And
in general, nonexperimental research is better when its authors attempt to justify the assumption that their proposed mediator
is uncorrelated with other variables, including unobserved variables, that may also be mediators. This sort of argument can
be made poorly or well. But even the best arguments of this type typically warrant far less confidence than arguments about
unconfoundedness that follow directly from manipulation of both the treatment and the mediator.
This discussion should make clear that the solution to bias in nonexperimental mediation analyses is unlikely to be another
nonexperimental mediation analysis. The problem is that factors affecting the mediator and the outcome are likely to covary.
We are not likely to solve this problem by controlling for more variables, measuring them more accurately, or applying newer
methods to nonexperimental data. To calculate unbiased estimates of mediation effects, we should look to experiments.
The simplest experimental design that permits accurate estimation of indirect effects entails direct manipulation of treatments
and mediators. We have described such a design elsewhere
(Bullock, Green, and Ha
2008), but in many cases, limited understanding of mediators precludes direct manipulation. For example, although we can assign
subjects to conditions in which their feelings of efficacy are likely to be heightened or diminished, we do not know how to
gain direct experimental control over efficacy. That is, we do not know how to assign specific levels of efficacy to different
subjects. The same is true of party identification, emotions, cultural norms, modes of information processing, and other likely
mediators of political processes. These variables and others are beyond direct experimental control.
But even when mediators are beyond direct experimental control, we can often manipulate them indirectly. The key in such cases
is to create an instrument for
M, the endogenous mediator. To be a valid instrument for
M, a variable must be correlated with
M but uncorrelated with
e3. Many variables are likely to satisfy the first condition: whatever
M is, it is usually not hard to think of a variable that is correlated with it, and once we have measured this new variable,
estimating the correlation is trivial. But satisfying the second condition is more difficult. Because
e3 is unobservable, we can never directly test whether it is uncorrelated with the potential instrument. Worse, almost every
variable that is correlated with
M is likely to be correlated with other factors that affect
Y, and thus likely to be correlated with
e3.
6
Fortunately, a familiar class of variables meets both conditions: assignment-to-treatment variables. Use of these instrumental
variables is especially common in analyses of field experiments, where compliance with the treatment is likely to be partial.
For example,
Gerber and Green (
2000) use a field experiment to study various means of increasing voter turnout. They cannot directly manipulate the treatments
of interest: they cannot compel their subjects
to read mail, answer phone calls, or speak to face-to-face canvassers. Instead, they use random assignments to these treatments
as instruments for the treatments themselves. Doing so permits them to recover accurate estimates of treatment effects even
though the treatments are beyond direct experimental control. (For elaboration of this point, see
Angrist, Imbens, and Rubin [1996] and Gerber's chapter in this volume.)
Although the instrumental variables approach is increasingly used to estimate average treatment effects, it has not yet been
used in political science to study mediation. We think that it should be. It has already been used multiple times to study
mediation in social psychology, and its use in that discipline suggests how it might be used in ours. For example,
Zanna and Cooper (
1974) hypothesize that attitude-behavior conflict produces feelings of unpleasant tension (“aversive arousal”), which in turn
produces attitude change. They cannot directly manipulate levels of tension, so they use an instrument to affect it indirectly:
subjects swallow a pill and are randomly assigned to hear that it will make them tense, make them relax, or have no effect.
In a related vein,
Bolger and Amarel (
2007) hypothesize that the effect of social support on the stress levels of recipients is mediated by efficacy: support reduces
recipients’ stress by raising their feelings of efficacy. Bolger and Amarel cannot directly assign different levels of efficacy
to different participants in their experiment. Instead, they randomly assign subjects to receive personal messages that are
designed to promote or diminish their feelings of efficacy. In this way, they indirectly manipulate efficacy.
To see how such instruments might be created and used in political science, consider research on issue framing. A controversial
hypothesis is that framing an issue in a particular way changes attitudes by increasing the accessibility of particular thoughts
about the issue, i.e., the ease with which particular thoughts come to mind
(see Iyengar and Kinder
1987, esp. ch. 7;
Nelson, Clawson, and Oxley
1997;
Miller and Krosnick
2000). Political scientists do not know how to directly manipulate the accessibility of particular thoughts, but they do know
how to indirectly manipulate accessibility by priming people in different ways
(e.g., Burdein, Lodge, and Taber
2006, esp. 363–64; see also Lodge and Taber's chapter in this volume). Experimental analysis of the hypothesis is therefore possible.
Following
Equation (3), consider the model:
In this model,
framing indicates whether subjects were assigned to a control condition (
framing = 0) or an issue frame (
framing = 1);
accessibility is reaction times in milliseconds in a task designed to gauge the accessibility of particular thoughts about the issue; and
e3 is a disturbance representing the cumulative effect of other variables. Crucially,
accessibility is not randomly assigned. It is likely to be affected by framing and to be correlated with unobserved variables represented
by
e3: age, intelligence, and political predispositions, among others.
The OLS estimator of b, the effect of accessibility, is therefore likely to be biased. (The OLS estimator of d, the direct effect of the framing manipulation, is also likely to be biased.) But suppose that in addition to the framing
manipulation and the measurement of accessibility, some subjects are randomly assigned to a condition in which relevant considerations
are primed. This priming manipulation may make certain thoughts about the issue more accessible. In this case, accessibility
remains nonexperimental, but the priming intervention generates an instrumental variable that we can use to consistently estimate
b. If we also estimate a – for example, by conducting a second experiment in which only framing is manipulated – our estimator of ab, the extent to which priming mediates framing, will also be consistent.
The most common objection to experimental mediation approaches is that they often cannot be used because mediators often cannot
be manipulated. We take up this objection later in this chapter, but
for the moment, we stress that researchers need not seek complete experimental control over mediators. They need only seek
some randomization-based purchase on mediators. Consider, for example, one of the best-known and least tractable variables
in political behavior research: party identification. The history of party ID studies suggests that it should be difficult
to manipulate. It is one of the most stable individual-level influences on votes and attitudes, and no one knows how to assign
different levels of party ID to different subjects. But party ID can be changed by experiments, and such experiments are the
key to understanding its mediating power. For example,
Brader and Tucker (
2008) use survey experiments to show that party cues can change Russians’ party IDs. And
Gerber, Huber, and Washington (2010) use a field experiment to show that registering with a party can produce long-term changes
in party ID. The most promising path to secure inferences about party ID as a mediator is to conduct studies in which interventions
like these are coupled with manipulations of policy preferences, candidate evaluations, or other treatments. And in general,
the most promising path to secure inferences about mediation is to design studies that include experimental manipulations
of both treatments and
mediators.
Despite its promise, the experimental approach has limitations that merit more attention than they typically receive. It requires
researchers to devise experimental manipulations that affect one mediator without affecting others.
Even if researchers succeed, their estimates of indirect effects will typically apply only to a subset of the experimental
sample. Finally, if causal effects are not identical for all members of a sample, then even a well-designed experiment may
lead to inaccurate inferences about indirect effects. We discuss these limitations at length in other work (Bullock, Green,
and Ha
2010; Green, Ha, and Bullock
2010); here, we offer a brief overview of each.
7
An experimental intervention is useful for mediation analysis if it affects one mediator without affecting others. If the
intervention instead affects more than one mediator, it violates the exclusion restriction – in terms of Equation (3), it
is correlated with
e3 – and is not a valid instrument. In this case, the instrumental variables estimator of the indirect effect will be biased.
For example, issue frames may affect attitudes not only by changing the accessibility of relevant considerations, but also
by changing the subjective relevance of certain values to the issue at hand (Nelson et al.
1997). In this case, an experimental intervention can identify the mediating role of accessibility only if it primes relevant
considerations without affecting the subjective relevance of different values. And by the same token, an experimental intervention
will identify the mediating role of value weighting only if it affects the subjective relevance of different values without
changing the accessibility of considerations. The general challenge for experimental researchers, then, is to devise manipulations
that affect one mediator without affecting others.
8
Even if researchers isolate particular mediators, they must confront another dilemma: some subjects never take a treatment
even if they are assigned to take it, and a treatment effect cannot be meaningfully estimated for such people. Consequently,
the experimental approach to mediation analysis produces estimates of the average treatment effect not for all subjects but
only for “compliers” who can be induced by random assignment to take it
(Imbens and Angrist
1994). For example, if some subjects are assigned to watch a presidential campaign advertisement while others are assigned to
a no-advertisement control group, then the average effect of the ad can be identified not for all subjects but only for 1)
treatment-group subjects who are induced by random assignment to watch the ad, and (2) control-group subjects who would have
been induced to watch the ad if they had been assigned to the treatment group. One may assume that the average indirect effect
is the same for these subjects as for others, but this is an assumption, not an experimental result. Strictly speaking, estimates
of the average indirect effect apply only to a subset of the sample. We can usually learn something about the characteristics
of this subset
(Angrist and Pischke
2009, 166–72), but we can never know exactly which subjects belong to it.
An unintuitive consequence follows: even if we use experiments to manipulate both a treatment and a mediator, we may not be
able to estimate an average indirect effect for our experimental sample or any subset of it. To see why, recall that the indirect
effect of
X on
Y in Equations (1)–(3) is
ab. By manipulating
X, we can recover
â, an estimate of the average effect of
X on
M among those whose value of
X can be affected by the
X manipulation. And by manipulating
M, we can recover
b̂, an estimate of the average effect of
M on
Y among those whose value of
M can be affected by the
M manipulation. If these two populations are the same,
âb̂ is a sensible estimate of the local average treatment effect. But if these two populations differ – if one set of subjects
is affected by the manipulation of
X but a different set is affected by the manipulation of
M –
âb̂ is the causal effect of
X on
M for one group of people times the causal effect of
M on
Y for another group of people. This product has no causal interpretation. It is just an unclear mixture of causal effects for
different groups.
9
A related problem is that experiments cannot lead to accurate estimates of indirect effects when the effects of
X on
M are not the same for all subjects or when the effects of
M on
Y are not the same for all subjects. When we are not studying mediation, the assumption of unvarying effects does little harm:
if the effect of randomly manipulated
X on
Y varies across subjects, and we regress
Y on
X, then the coefficient on
X simply indicates the average effect of
X. But if the effects of
X and
M vary across subjects, it will typically be difficult to estimate an average indirect effect
(Glynn
2010). To see why, consider an experimental sample in which there are two groups of subjects. In the first group, the effect of
X on
M is positive, and the effect of
M on
Y is also positive. In the second group, the effect of
X on
M is negative, and the effect of
M on
Y is also negative. In this case, the indirect effect of
X is positive for every subject in the sample: to slightly adapt the notation of Equations (1) and (3),
aibi is positive for every subject. But
â, the estimate of the average effect of
X on
M, may be positive, negative, or zero. And
b̂, the estimate of the average effect of
M on
Y, may be positive, negative, or zero. As a result, the estimate of the average indirect effect,
âb̂, may be zero or negative – even though the true indirect effect is positive for every subject.
Such problems may arise whenever different people are affected in different ways by
X and
M. For example,
Cohen (
2003) wants to understand how reference-group cues (
X) affect attitudes toward social policy (
Y). In his experiments, politically conservative subjects receive information about a generous welfare policy; some of these
subjects are told that the policy is endorsed by the Republican Party, while others receive no endorsement
information. Cohen's findings are consistent with cues (endorsements) promoting systematic thinking (
M) about the policy information, and with systematic thinking in turn promoting positive attitudes toward the policy
(Cohen
2003, esp. 817).
10 On the other hand,
Petty and Wegener (
1998, 345) and others suggest that reference-group cues inhibit systematic thinking about information, and that such thinking
promotes the influence of policy details – which might be expected to lead, in this case, to negative attitudes toward the
welfare policy among the conservative subjects. For present purposes, there is no need to favor either of these theories or
to attempt a reconciliation. We need only note that they suggest a case in which causal effects may be heterogeneous, and
in which mediation analysis is therefore difficult. Let some subjects in an experiment be “Cohens”: for these people, exposure
to reference group cues heightens systematic thinking (
ai is positive), and systematic thinking makes attitudes toward a generous welfare policy more favorable (
bi is positive). But other subjects are “Petties”: for them, exposure to reference group cues limits systematic thinking (
ai is negative), and systematic thinking makes attitudes toward a generous welfare policy less favorable (
bi is negative). Here again, the indirect effect is positive for every subject because
aibi > 0 for all
i. But if the experimental sample includes both Cohens and Petties,
â and
b̂ may each be positive, negative, or zero. Conventional estimates of the average indirect effect –
âb̂ and related quantities – may therefore be zero or even negative.
Moreover, causal effects need not differ so sharply across members of a sample to make mediation analysis problematic. Conventional
estimates of indirect effects will be biased if
a and
b merely covary among subjects within a sample. For example, if a subset of subjects is more sensitive than the rest of the
sample to changes in
X and to changes in
M, estimates of indirect effects will be biased. This problem cannot be traced to a deficiency in the methods that are often
used to calculate indirect effects: it is fundamental, not a matter of statistical
technique
(Robins
2003;
Glynn
2010).
These limitations of experimental mediation analysis – it requires experimenters to isolate particular mediators, produces
estimates that apply only to an unknown subset of subjects, and cannot produce meaningful inferences about mediation when
causal effects covary within a sample – are daunting. Experiments are often seen as simplifying causal inference, but taken
together, these limitations imply that strong inferences about mediation are likely to be difficult even when researchers
use fully experimental methods of mediation analysis. Still, none of our cautions implies that experiments are useless for
mediation analysis. Nor do they imply that experimental mediation analysis is no better than the nonexperimental alternative. Unlike nonexperimental methods, experiments offer – albeit under limited circumstances – a systematic
way to identify mediation effects. And the limitations that we describe are helpful inasmuch as they delineate an agenda for
future mediation analysis.
First, researchers who do not manipulate mediators should try to explain why the mediators are independent of the disturbances in
their regression equations – after all, the accuracy of their estimates hinges on this assumption. In practice, justifying
this assumption entails describing unmeasured mediators that may link X to Y and explaining why these mediators do not covary with the measured mediators. Such efforts are rarely undertaken, but without
them it is hard to hold out hope that nonexperimental mediation analysis will generate credible findings about mediation.
Second, researchers who experimentally manipulate mediators should explain why they believe that their manipulations are
isolating individual mediators. This entails describing the causal paths by which
X may affect
Y and explaining why each experimental manipulation affects only one of these paths. The list of alternative causal paths may
be extensive, and multiple experiments may be needed to demonstrate that a given intervention tends not to affect the alternative
paths in question.
Third, researchers can improve the state of mediation analysis simply by manipulating treatments and then measuring the effects
of their manipulations on many different outcomes. To see how this can improve mediation analysis, consider studies of the effects of campaign contact on voter turnout. In
addition to assessing whether a particular kind of contact increases turnout, one might also survey participants to determine
whether this kind of contact affects interest in politics, feelings of civic responsibility, knowledge about where and how
to vote, and other potential mediators. In a survey or laboratory experiment, this extra step need not entail a new survey:
relevant questions can instead be added to the post-test questionnaire. Because this kind of study does not include manipulations
of both treatments and mediators, it cannot reliably identify mediation effects. But if some variables seem to be unaffected
by the treatment, one may begin to argue that they do not explain why the treatment works.
Fourth, researchers should know that if the effects of
X and
M vary from subject to subject within a sample, it may be impossible to estimate the average indirect effect for the entire
sample. To determine whether this is a problem, one can examine the effects of
X and
M among different types of subjects. If the effects differ little from group to group (e.g., from men to women, whites to nonwhites,
the wealthy to the poor), we can be relatively confident that causal heterogeneity is not affecting our analysis.
11 In contrast, if there are large between-group differences in the effects of
X or
M, then mediation estimates made for an entire sample may be inaccurate even if
X and
M have been experimentally manipulated. In this case, researchers should aim to make multiple inferences for relatively homogeneous
subgroups rather than single inferences about the size of an indirect effect for an entire
sample.
Angrist, Joshua D., Guido W. Imbens, and Donald B. Rubin. 1996. “Identification of Causal Effects Using Instrumental Variables.” Journal of the American Statistical Association 91: 444–55.
Angrist, Joshua D., Victor Lavy, and Analia Schlosser. 2010. “Multiple Experiments for the Causal Link between the Quantity and Quality of Children.” Journal of Labor Economics 28: 773–824.
Angrist, Joshua D., and Jörn-Steffen Pischke. 2009. Mostly Harmless Econometrics: An Empiricist's Companion. Princeton, NJ: Princeton University Press.
Baron, Reuben M., and David A. Kenny. 1986. “The Moderator-Mediator Variable Distinction in Social Psychological Research: Conceptual, Strategic, and Statistical Considerations.” Journal of Personality and Social Psychology 51: 1173–82.
Bartels, Larry M. 1991. “Instrumental and ‘Quasi-Instrumental’ Variables.” American Journal of Political Science 35: 777–800.
Bolger, Niall, and David Amarel. 2007. “Effects of Social Support Visibility on Adjustment to Stress: Experimental Evidence.” Journal of Personality and Social Psychology 92: 458–75.
Bound, John, David A. Jaeger, and Regina M. Baker. 1995. “Problems with Instrumental Variables Estimation When the Correlation between the Instruments and the Endogenous Explanatory
Variable Is Weak.” Journal of the American Statistical Association 90: 443–50.
Brader, Ted A., and Joshua A. Tucker. 2008. “Reflective and Unreflective Partisans? Experimental Evidence on the Links between Information, Opinion, and Party Identification.”
Manuscript, New York University.
Brader, Ted, Nicholas A. Valentino, and Elizabeth Suhay. 2008. “What Triggers Public Opposition to Immigration? Anxiety, Group Cues, and Immigration Threat.” American Journal of Political Science 52: 959–78.
Bullock, John G., Donald P. Green, and Shang E. Ha. 2008. “Experimental Approaches to Mediation: A New Guide for Assessing Causal Pathways.” Unpublished manuscript, Yale University.
Bullock, John G., Donald P. Green, and Shang E. Ha. 2010. “Yes, But What's the Mechanism? (Don't Expect an Easy Answer).” Journal of Personality and Social Psychology 98: 550–58.
Burdein, Inna, Milton Lodge, and Charles Taber. 2006. “Experiments on the Automaticity of Political Beliefs and Attitudes.” Political Psychology 27: 359–71.
Campbell, Angus, Philip E. Converse, Warren Miller, and Donald Stokes. 1960. The American Voter. Chicago: The University of Chicago Press.
Clarke, Kevin A. 2009. “Return of the Phantom Menace: Omitted Variable Bias in Political Research.” Conflict Management and Peace Science 26: 46–66.
Cohen, Geoffrey L. 2003. “Party over Policy: The Dominating Impact of Group Influence on Political Beliefs.” Journal of Personality and Social Psychology 85: 808–22.
Downs, Anthony. 1957. An Economic Theory of Democracy. New York: HarperCollins.
Fowler, James H., and Christopher T. Dawes. 2008. “Two Genes Predict Voter Turnout.” Journal of Politics 70: 579–94.
Frangakis, Constantine E., and Donald B. Rubin. 2002. “Principal Stratification in Causal Inference.” Biometrics 58: 21–29.
Gelman, Andrew, and Jennifer Hill. 2007. Data Analysis Using Regression and Multilevel Hierarchical Models. New York: Cambridge University Press.
Gerber, Alan S., and Donald P. Green. 2000. “The Effects of Canvassing, Telephone Calls, and Direct Mail on Voter Turnout: A Field Experiment.” American Political Science Review 94: 653–62.
Gerber, Alan S., Gregory A. Huber, and Ebonya Washington. 2010. “Party Affiliation, Partisanship, and Political Beliefs: A Field Experiment.” American Political Science Review 104: 720–44.
Glynn, Adam N. 2010. “The Product and Difference Fallacies for Indirect Effects.” Unpublished manuscript, Harvard University.
Green, Donald P., Shang E. Ha, and John G. Bullock. 2010. “Enough Already about ‘Black Box’ Experiments: Studying Mediation Is More Difficult Than Most Scholars Suppose.” Annals of the American Academy of Political and Social Sciences 628: 200–8.
Imai, Kosuke,
Luke Keele,
Dustin Tingley, and
Teppei Yamamoto.
2010. “Unpacking the Black Box: Learning about Causal Mechanisms from Experimental and Observational Studies.” Unpublished manuscript,
Princeton University. Retrieved from
http://imai.princeton.edu/research/files/mediationP.pdf (November 21, 2010).
Imai, Kosuke, Luke Keele, and Teppei Yamamoto. 2010. “Identification, Inference, and Sensitivity Analysis for Causal Mediation Effects.” Statistical Science 25: 51–71.
Imbens, Guido W., and Joshua D. Angrist. 1994. “Identification and Estimation of Local Average Treatment Effects.” Econometrica 62: 467–75.
Iyengar, Shanto, and Donald R. Kinder. 1987. News That Matters: Television and American Opinion. Chicago: The University of Chicago Press.
James, Lawrence R. 1980. “The Unmeasured Variables Problem in Path Analysis.” Journal of Applied Psychology 65: 415–21.
James, Lawrence R. 2008. “On the Path to Mediation.” Organizational Research Methods 11: 359–63.
Judd, Charles M., and David A. Kenny. 1981. “Process Analysis: Estimating Mediation in Treatment Evaluations.” Evaluation Review 5: 602–19.
Kenny, David A. 2008. “Reflections on Mediation.” Organizational Research Methods 11: 353–58.
King, Gary, and Langche Zeng. 2006. “The Dangers of Extreme Counterfactuals.” Political Analysis 14: 131–59.
LaLonde, Robert J. 1986. “Evaluating the Econometric Evaluations of Training Programs with Experimental Data.” American Economic Review 76: 604–20.
MacKinnon, David P., Chondra M. Lockwood, Jeanne M. Hoffman, Stephen G. West, and Virgil Sheets. 2002. “A Comparison of Methods to Test Mediation and Other Intervening Variable Effects.” Psychological Methods 7: 83–104.
Malhotra, Neil, and Jon A. Krosnick. 2007. “Retrospective and Prospective Performance Assessments during the 2004 Election Campaign: Tests of Mediation and News Media
Priming.” Political Behavior 29: 249–78.
Miller, Joanne M., and Jon A. Krosnick. 2000. “News Media Impact on the Ingredients of Presidential Evaluations.” American Journal of Political Science 44: 295–309.
Morgan, Stephen L., and Christopher Winship. 2007. Counterfactuals and Causal Inference. New York: Cambridge University Press.
Nelson, Thomas E. 2004. “Policy Goals, Public Rhetoric, and Political Attitudes.” Journal of Politics 66: 581–605.
Nelson, Thomas E., Rosalee A. Clawson, and Zoe M. Oxley. 1997. “Media Framing of a Civil Liberties Conflict and Its Effect on Tolerance.” American Political Science Review 91: 567–94.
Pearl, Judea.
2010. “The Mediation Formula: A Guide to the Assessment of Causal Pathways in Non-Linear Models.” Unpublished manuscript, University
of California, Los Angeles. Retrieved from
http://ftp.cs.ucla.edu/~kaoru/r363.pdf (November 21, 2010).
Petty, Richard E., and Duane T. Wegener. 1998. “Attitude Change: Multiple Roles for Persuasion Variables.” In The Handbook of Social Psychology. vol. 1, 4th ed., eds. Daniel T. Gilbert, Susan T. Fiske, and Gardner Lindzey. New York: McGraw-Hill, 323–90.
Quiñones-Vidal, Elena, Juan J. López-Garcia, Maria Peñaranda-Ortega, and Francisco Tortosa-Gil. 2004. “The Nature of Social and Personality Psychology as Reflected in JPSP, 1965–2000.” Journal of Personality and Social Psychology 86: 435–52.
Robins, James M. 2003. “Semantics of Causal DAG Models and the Identification of Direct and Indirect Effects.” In Highly Structured Stochastic Systems, eds. Peter J. Green, Nils Lid Hjort, and Sylvia Richardson. New York: Oxford University Press, 70–81.
Rosenbaum, Paul R. 1984. “The Consequences of Adjustment for a Concomitant Variable That Has Been Affected by the Treatment.” Journal of the Royal Statistical Society, Series A 147: 656–66.
Rubin, Donald B. 2004. “Direct and Indirect Causal Effects via Potential Outcomes.” Scandinavian Journal of Statistics 31: 161–70.
Spencer, Steven J., Mark P. Zanna, and Geoffrey T. Fong. 2005. “Establishing a Causal Chain: Why Experiments Are Often More Effective Than Mediational Analyses in Examining Psychological
Processes.” Journal of Personality and Social Psychology 89: 845–51.
Zanna, Mark P., and Joel Cooper. 1974. “Dissonance and the Pill: An Attribution Approach to Studying the Arousal Properties of Dissonance.” Journal of Personality and Social Psychology 29: 703–9.