In statistics, a mediation model seeks to identify and explain the mechanism or process that underlies an observed relationship between an independent variable and a dependent variable via the inclusion of a third hypothetical variable, known as a mediator variable (also a mediating variable, intermediary variable, or intervening variable).[1] Rather than a direct causal relationship between the independent variable and the dependent variable, which is often false, a mediation model proposes that the independent variable influences the mediator variable, which in turn influences the dependent variable. Thus, the mediator variable serves to clarify the nature of the relationship between the independent and dependent variables.[2] [3]
Mediation analyses are employed to understand a known relationship by exploring the underlying mechanism or process by which one variable influences another variable through a mediator variable.[4] In particular, mediation analysis can contribute to better understanding the relationship between an independent variable and a dependent variable when these variables do not have an obvious direct connection.
Baron and Kenny (1986) laid out several requirements that must be met to form a true mediation relationship.[5] They are outlined below using a real-world example. See the diagram above for a visual representation of the overall mediating relationship to be explained. The original steps are as follows.
Relationship Duration
Regress the dependent variable on the independent variable to confirm that the independent variable is a significant predictor of the dependent variable.
Independent variable
\to
Y=\beta10+\beta11X+\varepsilon1
Regress the mediator on the independent variable to confirm that the independent variable is a significant predictor of the mediator. If the mediator is not associated with the independent variable, then it couldn’t possibly mediate anything.
Independent variable
\to
Me=\beta20+\beta21X+\varepsilon2
Regress the dependent variable on both the mediator and independent variable to confirm that a) the mediator is a significant predictor of the dependent variable, and b) the strength of the coefficient of the previously significant independent variable in Step #1 is now greatly reduced, if not rendered nonsignificant.
Y=\beta30+\beta31X+\beta32Me+\varepsilon3
The following example, drawn from Howell (2009),[6] explains each step of Baron and Kenny's requirements to understand further how a mediation effect is characterized. Step 1 and step 2 use simple regression analysis, whereas step 3 uses multiple regression analysis.
Such findings would lead to the conclusion implying that your feelings of competence and self-esteem mediate the relationship between how you were parented and how confident you feel about parenting your own children.
If step 1 does not yield a significant result, one may still have grounds to move to step 2. Sometimes there is actually a significant relationship between independent and dependent variables but because of small sample sizes, or other extraneous factors, there could not be enough power to predict the effect that actually exists.[7]
In the diagram shown above, the indirect effect is the product of path coefficients "A" and "B". The direct effect is the coefficient " C' ". The direct effect measures the extent to which the dependent variable changes when the independent variable increases by one unit and the mediator variable remains unaltered. In contrast, the indirect effect measures the extent to which the dependent variable changes when the independent variable is held constant and the mediator variable changes by the amount it would have changed had the independent variable increased by one unit.In linear systems, the total effect is equal to the sum of the direct and indirect (C' + AB in the model above). In nonlinear models, the total effect is not generally equal to the sum of the direct and indirect effects, but to a modified combination of the two.
A mediator variable can either account for all or some of the observed relationship between two variables.
Maximum evidence for mediation, also called full mediation, would occur if inclusion of the mediation variable drops the relationship between the independent variable and dependent variable (see pathway c′ in diagram above) to zero.
Partial mediation maintains that the mediating variable accounts for some, but not all, of the relationship between the independent variable and dependent variable. Partial mediation implies that there is not only a significant relationship between the mediator and the dependent variable, but also some direct relationship between the independent and dependent variable.
In order for either full or partial mediation to be established, the reduction in variance explained by the independent variable must be significant as determined by one of several tests, such as the Sobel test.[8] The effect of an independent variable on the dependent variable can become nonsignificant when the mediator is introduced simply because a trivial amount of variance is explained (i.e., not true mediation). Thus, it is imperative to show a significant reduction in variance explained by the independent variable before asserting either full or partial mediation. It is possible to have statistically significant indirect effects in the absence of a total effect.[9] This can be explained by the presence of several mediating paths that cancel each other out, and become noticeable when one of the cancelling mediators is controlled for. This implies that the terms 'partial' and 'full' mediation should always be interpreted relative to the set of variables that are present in the model. In all cases, the operation of "fixing a variable" must be distinguished from that of "controlling for a variable," which has been inappropriately used in the literature. The former stands for physically fixing, while the latter stands for conditioning on, adjusting for, or adding to the regression model. The two notions coincide only when all error terms (not shown in the diagram) are statistically uncorrelated. When errors are correlated, adjustments must be made to neutralize those correlations before embarking on mediation analysis (see Bayesian network).
See main article: Sobel test.
Sobel's test[8] is performed to determine if the relationship between the independent variable and dependent variable has been significantly reduced after inclusion of the mediator variable. In other words, this test assesses whether a mediation effect is significant. It examines the relationship between the independent variable and the dependent variable compared to the relationship between the independent variable and dependent variable including the mediation factor.
The Sobel test is more accurate than the Baron and Kenny steps explained above; however, it does have low statistical power. As such, large sample sizes are required in order to have sufficient power to detect significant effects. This is because the key assumption of Sobel's test is the assumption of normality. Because Sobel's test evaluates a given sample on the normal distribution, small sample sizes and skewness of the sampling distribution can be problematic (see Normal distribution for more details). Thus, the rule of thumb as suggested by MacKinnon et al., (2002) [10] is that a sample size of 1000 is required to detect a small effect, a sample size of 100 is sufficient in detecting a medium effect, and a sample size of 50 is required to detect a large effect.
The equation for Sobel is:[11]
z=
ab | ||||||||||||||||||
|
The bootstrapping method provides some advantages to the Sobel's test, primarily an increase in power. The Preacher and Hayes bootstrapping method is a non-parametric test and does not impose the assumption of normality. Therefore, if the raw data is available, the bootstrap method is recommended.[11] Bootstrapping involves repeatedly randomly sampling observations with replacement from the data set to compute the desired statistic in each resample. Computing over hundreds, or thousands, of bootstrap resamples provide an approximation of the sampling distribution of the statistic of interest. The Preacher–Hayes method provides point estimates and confidence intervals by which one can assess the significance or nonsignificance of a mediation effect. Point estimates reveal the mean over the number of bootstrapped samples and if zero does not fall between the resulting confidence intervals of the bootstrapping method, one can confidently conclude that there is a significant mediation effect to report.
As outlined above, there are a few different options one can choose from to evaluate a mediation model.
Bootstrapping[12] [13] is becoming the most popular method of testing mediation because it does not require the normality assumption to be met, and because it can be effectively utilized with smaller sample sizes (N < 25). However, mediation continues to be most frequently determined using the logic of Baron and Kenny[14] or the Sobel test. It is becoming increasingly more difficult to publish tests of mediation based purely on the Baron and Kenny method or tests that make distributional assumptions such as the Sobel test. Thus, it is important to consider your options when choosing which test to conduct.
While the concept of mediation as defined within psychology is theoretically appealing, the methods used to study mediation empirically have been challenged by statisticians and epidemiologists[15] [16] [17] and interpreted formally.[18]
Hayes (2009) critiqued Baron and Kenny's mediation steps approach, and as of 2019, David A. Kenny on his website stated that mediation can exist in the absence of a 'significant' total effect (sometimes referred to as "inconsistent mediation"), and therefore step 1 of the original 1986 approach may not be needed. Later publications by Hayes questioned the concepts of full mediation and partial mediation, and advocated for the abandonment of these terms and of the steps in classical (1986) mediation.
Experimental approaches to mediation must be carried out with caution. First, it is important to have strong theoretical support for the exploratory investigation of a potential mediating variable. A criticism of a mediation approach rests on the ability to manipulate and measure a mediating variable. Thus, one must be able to manipulate the proposed mediator in an acceptable and ethical fashion. As such, one must be able to measure the intervening process without interfering with the outcome. The mediator must also be able to establish construct validity of manipulation. One of the most common criticisms of the measurement-of-mediation approach is that it is ultimately a correlational design. Consequently, it is possible that some other third variable, independent from the proposed mediator, could be responsible for the proposed effect. However, researchers have worked hard to provide counter-evidence to this disparagement. Specifically, the following counter-arguments have been put forward:[4]
Mediation can be an extremely useful and powerful statistical test; however, it must be used properly. It is important that the measures used to assess the mediator and the dependent variable are theoretically distinct and that the independent variable and mediator cannot interact. Should there be an interaction between the independent variable and the mediator one would have grounds to investigate moderation.
See main article: Spurious relationship.
Another model that is often tested is one in which competing variables in the model are alternative potential mediators or an unmeasured cause of the dependent variable. An additional variable in a causal model may obscure or confound the relationship between the independent and dependent variables. Potential confounders are variables that may have a causal impact on both the independent variable and dependent variable. They include common sources of measurement error (as discussed above) as well as other influences shared by both the independent and dependent variables.
In experimental studies, there is a special concern about aspects of the experimental manipulation or setting that may account for study effects, rather than the motivating theoretical factor. Any of these problems may produce spurious relationships between the independent and dependent variables as measured. Ignoring a confounding variable may bias empirical estimates of the causal effect of the independent variable.
A suppressor variable increases the predictive validity of another variable when included in a regression equation. Suppression can occur when a single causal variable is related to an outcome variable through two separate mediator variables, and when one of those mediated effects is positive and one is negative. In such a case, each mediator variable suppresses or conceals the effect that is carried through the other mediator variable. For example, higher intelligence scores (a causal variable, A) may cause an increase in error detection (a mediator variable, B) which in turn may cause a decrease in errors made at work on an assembly line (an outcome variable, X); at the same time, intelligence could also cause an increase in boredom (C), which in turn may cause an increase in errors (X). Thus, in one causal path intelligence decreases errors, and in the other it increases them. When neither mediator is included in the analysis, intelligence appears to have no effect or a weak effect on errors. However, when boredom is controlled intelligence will appear to decrease errors, and when error detection is controlled intelligence will appear to increase errors. If intelligence could be increased while only boredom was held constant, errors would decrease; if intelligence could be increased while holding only error detection constant, errors would increase.
In general, the omission of suppressors or confounders will lead to either an underestimation or an overestimation of the effect of A on X, thereby either reducing or artificially inflating the magnitude of a relationship between two variables.
Other important third variables are moderators. Moderators are variables that can make the relationship between two variables either stronger or weaker. Such variables further characterize interactions in regression by affecting the direction and/or strength of the relationship between X and Y. A moderating relationship can be thought of as an interaction. It occurs when the relationship between variables A and B depends on the level of C. See moderation for further discussion.
Mediation and moderation can co-occur in statistical models. It is possible to mediate moderation and moderate mediation.
Moderated mediation is when the effect of the treatment A on the mediator and/or the partial effect B on the dependent variable depend in turn on levels of another variable (moderator). Essentially, in moderated mediation, mediation is first established, and then one investigates if the mediation effect that describes the relationship between the independent variable and dependent variable is moderated by different levels of another variable (i.e., a moderator). This definition has been outlined by Muller, Judd, and Yzerbyt (2005)[20] and Preacher, Rucker, and Hayes (2007).[21]
There are five possible models of moderated mediation, as illustrated in the diagrams below.
In addition to the models mentioned above, a new variable can also exist which moderates the relationship between the independent variable and mediator (the A path) while at the same time have the new variable moderate the relationship between the independent variable and dependent variable (the C Path).
Mediated moderation is a variant of both moderation and mediation. This is where there is initially overall moderation and the direct effect of the moderator variable on the outcome is mediated. The main difference between mediated moderation and moderated mediation is that for the former there is initial (overall) moderation and this effect is mediated and for the latter there is no moderation but the effect of either the treatment on the mediator (path A) is moderated or the effect of the mediator on the outcome (path B) is moderated.
In order to establish mediated moderation, one must first establish moderation, meaning that the direction and/or the strength of the relationship between the independent and dependent variables (path C) differs depending on the level of a third variable (the moderator variable). Researchers next look for the presence of mediated moderation when they have a theoretical reason to believe that there is a fourth variable that acts as the mechanism or process that causes the relationship between the independent variable and the moderator (path A) or between the moderator and the dependent variable (path C).
The following is a published example of mediated moderation in psychological research.[22] Participants were presented with an initial stimulus (a prime) that made them think of morality or made them think of might. They then participated in the Prisoner's Dilemma Game (PDG), in which participants pretend that they and their partner in crime have been arrested, and they must decide whether to remain loyal to their partner or to compete with their partner and cooperate with the authorities. The researchers found that prosocial individuals were affected by the morality and might primes, whereas proself individuals were not. Thus, social value orientation (proself vs. prosocial) moderated the relationship between the prime (independent variable: morality vs. might) and the behaviour chosen in the PDG (dependent variable: competitive vs. cooperative).
The researchers next looked for the presence of a mediated moderation effect. Regression analyses revealed that the type of prime (morality vs. might) mediated the moderating relationship of participants’ social value orientation on PDG behaviour. Prosocial participants who experienced the morality prime expected their partner to cooperate with them, so they chose to cooperate themselves. Prosocial participants who experienced the might prime expected their partner to compete with them, which made them more likely to compete with their partner and cooperate with the authorities. In contrast, participants with a pro-self social value orientation always acted competitively.
Muller, Judd, and Yzerbyt (2005)[20] outline three fundamental models that underlie both moderated mediation and mediated moderation. Mo represents the moderator variable(s), Me represents the mediator variable(s), and εi represents the measurement error of each regression equation.
Moderation of the relationship between the independent variable (X) and the dependent variable (Y), also called the overall treatment effect (path C in the diagram).
Y=\beta40+\beta41X+\beta42Mo+\beta43XMo+\varepsilon4
Moderation of the relationship between the independent variable and the mediator (path A).
Me=\beta50+\beta51X+\beta52Mo+\beta53XMo+\varepsilon5
Moderation of both the relationship between the independent and dependent variables (path A) and the relationship between the mediator and the dependent variable (path B).
Y=\beta60+\beta61X+\beta62Mo+\beta63XMo+\beta64Me+\beta65MeMo+\varepsilon6
Mediation analysis quantifies the extent to which a variable participates in the transmittanceof change from a cause to its effect. It is inherently a causalnotion, hence it cannot be defined in statistical terms. Traditionally,however, the bulk of mediation analysis has been conductedwithin the confines of linear regression, with statistical terminology masking the causal character of therelationships involved. This led to difficulties,biases, and limitations that have been alleviated bymodern methods of causal analysis, based on causal diagramsand counterfactual logic.
The source of these difficulties lies in defining mediationin terms of changes induced by adding a third variables into a regression equation. Such statistical changes areepiphenomena which sometimes accompany mediation but, in general, fail to capture the causal relationships thatmediation analysis aims to quantify.
The basic premise of the causal approach is that it isnot always appropriate to "control" for the mediator Mwhen we seek to estimate the direct effect of X on Y(see the Figure above).The classical rationale for "controlling" for M" is that, if we succeed in preventing M from changing, then whatever changes we measure in Y are attributable solely to variations in X and we are justified then in proclaiming the effect observed as "direct effect of X on Y." Unfortunately,"controlling for M" does not physically prevent M from changing;it merely narrows the analyst's attention to casesof equal M values. Moreover, the language of probability theory does not possess the notation to express the idea of "preventing M from changing" or "physically holding M constant".The only operator probability provides is "Conditioning"which is what we do when we "control" for M, or add M as a regressor in the equation for Y. The result is that, instead of physically holding M constant (say at M = m) and comparing Y for units under X = 1' to those under X = 0, we allow M to vary but ignore all units except those in which M achieves the value M = m. These two operations are fundamentally different, and yield different results,[23] [24] except in the case of no omitted variables. Improperly conditioning mediated effects can be a type of bad control.
To illustrate, assume that the error terms of M and Yare correlated. Under such conditions, thestructural coefficient B and A (between M and Y and between Y and X)can no longer be estimated by regressing Y on X and M.In fact, the regression slopes may both be nonzero even when C is zero.[25] This has two consequences. First, new strategies must be devised for estimating the structural coefficients A, B and C. Second, the basic definitions of direct and indirect effects must go beyond regression analysis, and should invoke an operation that mimics "fixing M", rather than "conditioning on M."
Such an operator, denoted do(M = m), was defined in Pearl (1994) and it operates by removing the equation of M and replacing it by a constant m. For example, if the basic mediation model consists of the equations:
X=f(\varepsilon1),~~M=g(X,\varepsilon2),~~Y=h(X,M,\varepsilon3),
then after applying the operator do(M = m) the model becomes:
X=f(\varepsilon1),~~M=m,~~Y=h(X,m,\varepsilon3)
and after applying the operator do(X = x) the model becomes:
X=x,M=g(x,\varepsilon2),Y=h(x,M,\varepsilon3)
where the functions f and g, as well as thedistributions of the error terms ε1 and ε3 remainunaltered. If we further rename the variables M and Y resulting from do(X = x)as M(x) and Y(x), respectively, we obtain what came to be known as "potentialoutcomes"[26] or "structural counterfactuals".[27] These new variables provide convenient notationfor defining direct and indirect effects. In particular,four types of effects have been defined for thetransition from X = 0 to X = 1:
(a) Total effect –
TE=E[Y(1)-Y(0)]
CDE(m)=E[Y(1,m)-Y(0,m)]
NDE=E[Y(1,M(0))-Y(0,M(0))]
NIE=E[Y(0,M(1))-Y(0,M(0))]
These effects have the following interpretations:
A controlled version of the indirect effect does notexist because there is no way of disabling the direct effect by fixing a variable to a constant.
According to these definitions the total effect can be decomposed as a sum
TE=NDE-NIEr
The power of these definitions lies in their generality; they are applicable to models with arbitrary nonlinear interactions, arbitrary dependencies among the disturbances, and both continuous and categorical variables.
In linear analysis, all effects are determined by sums of products of structural coefficients, giving
\begin{align} TE&=C+AB\\ CDE(m)&=NDE=C,independentofm\\ NIE&=AB. \end{align}
\begin{align} TE&=E(Y\midX=1)-E(Y\midX=0)\\ CDE(m)&=E(Y\midX=1,M=m)-E(Y\midX=0,M=m)\\ NDE&=\summ[E(Y|X=1,M=m)-E(Y\midX=0,M=m)]P(M=m\midX=0)\\ NIE&=\summ[P(M=m\midX=1)-P(M=m\midX=0)]E(Y\midX=0,M=m). \end{align}
The last two equations are called Mediation Formulas[30] [31] [32] and have become the target of estimation in many studies of mediation.[28] They give distribution-free expressions for direct and indirect effects and demonstrate that, despite the arbitrary nature of the error distributions and the functions f, g, and h, mediated effects can nevertheless be estimated from data using regression. The analyses of moderated mediation and mediating moderators fall as special cases of the causal mediation analysis, and the mediation formulas identify how various interactions coefficients contribute to the necessary and sufficient components of mediation.
Assume the model takes the form
\begin{align} X&=\varepsilon1\\ M&=b0+b1X+\varepsilon2\\ Y&=c0+c1X+c2M+c3XM+\varepsilon3\end{align}
c3
TE
TE
b1c2/TE
(TE-c1)/TE
\begin{align} NDE&=c1+b0c3\\ NIE&=b1c2\\ TE&=c1+b0c3+b1(c2+c3)\\ &=NDE+NIE+b1c3. \end{align}
Thus, the fraction of output response for which mediation would be sufficient is
NIE | |
TE |
=
b1c2 | |
c1+b0c3+b1(c2+c3) |
,
while the fraction for which mediation would be necessary is
1-
NDE | |
TE |
=
b1(c2+c3) | |
c1+b0c3+b1(c2+c3) |
.
These fractions involve non-obvious combinationsof the model's parameters, and can be constructedmechanically with the help of the Mediation Formula. Significantly, due to interaction, a direct effect can be sustained even when the parameter
c1