In statistics, the principle of marginality is the fact that the average (or main) effects of variables in an analysis are marginal to their interaction effect—that is, the main effect of one explanatory variable captures the effect of that variable averaged over all values of a second explanatory variable whose value influences the first variable's effect. The principle of marginality implies that, in general, it is wrong to test, estimate, or interpret main effects of explanatory variables where the variables interact or, similarly, to model interaction effects but delete main effects that are marginal tothem.[1] While such models are interpretable, they lack applicability, as they ignore the dependence of a variable's effect upon another variable's value.
Nelder[2] and Venables[3] have argued strongly for the importance of this principle in regression analysis.
If two independent continuous variables, say x and z, both influence a dependent variable y, and if the extent of the effect of each independent variable depends on the level of the other independent variable then the regression equation can be written as:
yi=a+bxi+czi+d(xizi)+ei,
where i indexes observations, a is the intercept term, b, c, and d are effect size parameters to be estimated, and e is the error term.
If this is the correct model, then the omission of any of the right-side terms would be incorrect, resulting in misleading interpretation of the regression results.
With this model, the effect of x upon y is given by the partial derivative of y with respect to x; this is
b+dzi
zi
In addition:
drop1
, which does not drop the main effect terms from a model with interaction: "To my delight I see that marginality constraints between factor terms are by default honoured". In R, the marginality requirement of the dropterm
function (in package MASS) is stated in the Reference Manual.