Simultaneous equations models are a type of statistical model in which the dependent variables are functions of other dependent variables, rather than just independent variables.[1] This means some of the explanatory variables are jointly determined with the dependent variable, which in economics usually is the consequence of some underlying equilibrium mechanism. Take the typical supply and demand model: whilst typically one would determine the quantity supplied and demanded to be a function of the price set by the market, it is also possible for the reverse to be true, where producers observe the quantity that consumers demand and then set the price.[2]
Simultaneity poses challenges for the estimation of the statistical parameters of interest, because the Gauss–Markov assumption of strict exogeneity of the regressors is violated. And while it would be natural to estimate all simultaneous equations at once, this often leads to a computationally costly non-linear optimization problem even for the simplest system of linear equations.[3] This situation prompted the development, spearheaded by the Cowles Commission in the 1940s and 1950s,[4] of various techniques that estimate each equation in the model seriatim, most notably limited information maximum likelihood and two-stage least squares.[5]
Suppose there are m regression equations of the form
yit=y-i,t'\gammai+xit' \betai+uit, i=1,\ldots,m,
where i is the equation number, and is the observation index. In these equations xit is the ki×1 vector of exogenous variables, yit is the dependent variable, y−i,t is the ni×1 vector of all other endogenous variables which enter the ith equation on the right-hand side, and uit are the error terms. The “−i” notation indicates that the vector y−i,t may contain any of the y’s except for yit (since it is already present on the left-hand side). The regression coefficients βi and γi are of dimensions ki×1 and ni×1 correspondingly. Vertically stacking the T observations corresponding to the ith equation, we can write each equation in vector form as
yi=Y-i\gammai+Xi\betai+ui, i=1,\ldots,m,
Y\Gamma=X\Beta+U.
Postmultiplying the structural equation by, the system can be written in the reduced form as
Y=X\Beta\Gamma-1+U\Gamma-1=X\Pi+V.
Firstly, the rank of the matrix X of exogenous regressors must be equal to k, both in finite samples and in the limit as (this later requirement means that in the limit the expression should converge to a nondegenerate k×k matrix). Matrix Γ is also assumed to be non-degenerate.
Secondly, error terms are assumed to be serially independent and identically distributed. That is, if the tth row of matrix U is denoted by u(t), then the sequence of vectors should be iid, with zero mean and some covariance matrix Σ (which is unknown). In particular, this implies that, and .
Lastly, assumptions are required for identification.
The identification conditions require that the system of linear equations be solvable for the unknown parameters.
More specifically, the order condition, a necessary condition for identification, is that for each equation, which can be phrased as “the number of excluded exogenous variables is greater or equal to the number of included endogenous variables”.
The rank condition, a stronger condition which is necessary and sufficient, is that the rank of equals, where is a matrix which is obtained from by crossing out those columns which correspond to the excluded endogenous variables, and those rows which correspond to the included exogenous variables.
In simultaneous equations models, the most common method to achieve identification is by imposing within-equation parameter restrictions.[6] Yet, identification is also possible using cross equation restrictions.
To illustrate how cross equation restrictions can be used for identification, consider the following example from Wooldridge
\begin{align} y1&=\gamma12y2+\delta11z1+\delta12z2+\delta13z3+u1\\ y2&=\gamma21y1+\delta21z1+\delta22z2+u2 \end{align}
where z's are uncorrelated with u's and y's are endogenous variables. Without further restrictions, the first equation is not identified because there is no excluded exogenous variable. The second equation is just identified if, which is assumed to be true for the rest of discussion.
Now we impose the cross equation restriction of . Since the second equation is identified, we can treat as known for the purpose of identification. Then, the first equation becomes:
y1-\delta12z2=\gamma12y2+\delta11z1+\delta13z3+u1
Then, we can use as instruments to estimate the coefficients in the above equation since there are one endogenous variable and one excluded exogenous variable on the right hand side. Therefore, cross equation restrictions in place of within-equation restrictions can achieve identification.
The simplest and the most common estimation method for the simultaneous equations model is the so-called two-stage least squares method,[7] developed independently by and .[8] [9] [10] It is an equation-by-equation technique, where the endogenous regressors on the right-hand side of each equation are being instrumented with the regressors X from all other equations. The method is called “two-stage” because it conducts estimation in two steps:
Step 1: Regress Y−i on X and obtain the predicted values ;
Step 2: Estimate γi, βi by the ordinary least squares regression of yi on and Xi.
If the ith equation in the model is written as
yi=\begin{pmatrix}Y-i&Xi\end{pmatrix}\begin{pmatrix}\gammai\\\betai\end{pmatrix}+ui \equivZi\deltai+ui,
\hat\deltai=(\hat{Z}'i\hat{Z}
-1 | |
i) |
\hat{Z}'iyi =(Z'iPZi)-1Z'iPyi,
Indirect least squares is an approach in econometrics where the coefficients in a simultaneous equations model are estimated from the reduced form model using ordinary least squares.[11] [12] For this, the structural system of equations is transformed into the reduced form first. Once the coefficients are estimated the model is put back into the structural form.
The “limited information” maximum likelihood method was suggested by M. A. Girshick in 1947,[13] and formalized by T. W. Anderson and H. Rubin in 1949.[14] It is used when one is interested in estimating a single structural equation at a time (hence its name of limited information), say for observation i:
yi=Y-i\gammai+Xi\betai+ui\equivZi\deltai+ui
The structural equations for the remaining endogenous variables Y−i are not specified, and they are given in their reduced form:
Y-i=X\Pi+U-i
Notation in this context is different than for the simple IV case. One has:
Y-i
X-i
X
Z
The explicit formula for the LIML is:[15]
\hat\deltai=(Z'i(I-λ
-1 | |
M)Z | |
i) |
Z'i(I-λM)yi,
(\begin{bmatrix}yi\\Y-i\end{bmatrix}Mi\begin{bmatrix}yi&Y-i\end{bmatrix}) (\begin{bmatrix}yi\\Y-i\end{bmatrix}M\begin{bmatrix}yi&Y-i\end{bmatrix})-1
where, in a similar way, .
In other words, λ is the smallest solution of the generalized eigenvalue problem, see :
|\begin{bmatrix}yi&Y-i\end{bmatrix}'Mi\begin{bmatrix}yi&Y-i\end{bmatrix}-λ \begin{bmatrix}yi&Y-i\end{bmatrix}'M\begin{bmatrix}yi&Y-i\end{bmatrix}|=0
The LIML is a special case of the K-class estimators:[16]
\hat\delta=(Z'(I-\kappaM)Z)-1Z'(I-\kappaM)y,
with:
\delta=\begin{bmatrix}\betai&\gammai\end{bmatrix}
Z=\begin{bmatrix}Xi&Y-i\end{bmatrix}
I-\kappaM=I-M=P
The three-stage least squares estimator was introduced by .[18] [19] It can be seen as a special case of multi-equation GMM where the set of instrumental variables is common to all equations.[20] If all regressors are in fact predetermined, then 3SLS reduces to seemingly unrelated regressions (SUR). Thus it may also be seen as a combination of two-stage least squares (2SLS) with SUR.
Across fields and disciplines simultaneous equation models are applied to various observational phenomena. These equations are applied when phenomena are assumed to be reciprocally causal. The classic example is supply and demand in economics. In other disciplines there are examples such as candidate evaluations and party identification[21] or public opinion and social policy in political science;[22] [23] road investment and travel demand in geography;[24] and educational attainment and parenthood entry in sociology or demography.[25] The simultaneous equation model requires a theory of reciprocal causality that includes special features if the causal effects are to be estimated as simultaneous feedback as opposed to one-sided 'blocks' of an equation where a researcher is interested in the causal effect of X on Y while holding the causal effect of Y on X constant, or when the researcher knows the exact amount of time it takes for each causal effect to take place, i.e., the length of the causal lags. Instead of lagged effects, simultaneous feedback means estimating the simultaneous and perpetual impact of X and Y on each other. This requires a theory that causal effects are simultaneous in time, or so complex that they appear to behave simultaneously; a common example are the moods of roommates.[26] To estimate simultaneous feedback models a theory of equilibrium is also necessary – that X and Y are in relatively steady states or are part of a system (society, market, classroom) that is in a relatively stable state.[27]