Linear recurrence with constant coefficients explained

In mathematics (including combinatorics, linear algebra, and dynamical systems), a linear recurrence with constant coefficients[1] [2] (also known as a linear recurrence relation or linear difference equation) sets equal to 0 a polynomial that is linear in the various iterates of a variable—that is, in the values of the elements of a sequence. The polynomial's linearity means that each of its terms has degree 0 or 1. A linear recurrence denotes the evolution of some variable over time, with the current time period or discrete moment in time denoted as, one period earlier denoted as, one period later as, etc.

The solution of such an equation is a function of, and not of any iterate values, giving the value of the iterate at any time. To find the solution it is necessary to know the specific values (known as initial conditions) of of the iterates, and normally these are the iterates that are oldest. The equation or its variable is said to be stable if from any set of initial conditions the variable's limit as time goes to infinity exists; this limit is called the steady state.

Difference equations are used in a variety of contexts, such as in economics to model the evolution through time of variables such as gross domestic product, the inflation rate, the exchange rate, etc. They are used in modeling such time series because values of these variables are only measured at discrete intervals. In econometric applications, linear difference equations are modeled with stochastic terms in the form of autoregressive (AR) models and in models such as vector autoregression (VAR) and autoregressive moving average (ARMA) models that combine AR with other features.

Definitions

A linear recurrence with constant coefficients is an equation of the following form, written in terms of parameters and :

y_t = a_1y_ + \cdots + a_ny_ + b,

or equivalently as

y_= a_1y_ + \cdots + a_ny_t + b.

The positive integer

n

is called the order of the recurrence and denotes the longest time lag between iterates. The equation is called homogeneous if and nonhomogeneous if .

If the equation is homogeneous, the coefficients determine the characteristic polynomial (also "auxiliary polynomial" or "companion polynomial")

p(\lambda)= \lambda^n - a_1\lambda^ - a_2\lambda^-\cdots-a_

whose roots play a crucial role in finding and understanding the sequences satisfying the recurrence.

Conversion to homogeneous form

If, the equation

y_t = a_1y_ + \cdots + a_ny_ + b

is said to be nonhomogeneous. To solve this equation it is convenient to convert it to homogeneous form, with no constant term. This is done by first finding the equation's steady state value—a value such that, if successive iterates all had this value, so would all future values. This value is found by setting all values of equal to in the difference equation, and solving, thus obtaining

y^* = \frac

assuming the denominator is not 0. If it is zero, the steady state does not exist.

Given the steady state, the difference equation can be rewritten in terms of deviations of the iterates from the steady state, as

\left(y_t -y^*\right)= a_1\left(y_-y^*\right) + \cdots + a_n\left(y_-y^*\right)

which has no constant term, and which can be written more succinctly as

x_t= a_1x_ + \cdots + a_nx_

where equals . This is the homogeneous form.

If there is no steady state, the difference equation

y_t = a_1y_ + \cdots + a_ny_ + b

can be combined with its equivalent form

y_= a_1y_+ \cdots + a_ny_+ b

to obtain (by solving both for)

y_t - a_1y_ - \cdots - a_ny_ = y_- a_1y_- \cdots - a_ny_

in which like terms can be combined to give a homogeneous equation of one order higher than the original.

Solution example for small orders

The roots of the characteristic polynomial play a crucial role in finding and understanding the sequences satisfying the recurrence. If there are

d

distinct roots

r1,r2,\ldots,rd,

then each solution to the recurrence takes the forma_n = k_1 r_1^n + k_2 r_2^n + \cdots + k_d r_d^n,where the coefficients

ki

are determined in order to fit the initial conditions of the recurrence. When the same roots occur multiple times, the terms in this formula corresponding to the second and later occurrences of the same root are multiplied by increasing powers of

n

. For instance, if the characteristic polynomial can be factored as

(x-r)3

, with the same root

r

occurring three times, then the solution would take the forma_n = k_1 r^n + k_2 n r^n + k_3 n^2 r^n.[3]

Order 1

For order 1, the recurrencea_=r a_has the solution

an=rn

with

a0=1

and the most general solution is

an=krn

with

a0=k

. The characteristic polynomial equated to zero (the characteristic equation) is simply

t-r=0

.

Order 2

Solutions to such recurrence relations of higher order are found by systematic means, often using the fact that

an=rn

is a solution for the recurrence exactly when

t=r

is a root of the characteristic polynomial. This can be approached directly or using generating functions (formal power series) or matrices.

Consider, for example, a recurrence relation of the forma_ = Aa_+Ba_.

When does it have a solution of the same general form as

an=rn

? Substituting this guess (ansatz) in the recurrence relation, we find thatr^=Ar^+Br^must be true for all

n>1

.

Dividing through by

rn-2

, we get that all these equations reduce to the same thing:

\begin r^2 &= Ar + B, \\ r^2 - Ar - B &= 0, \end

which is the characteristic equation of the recurrence relation. Solve for

r

to obtain the two roots

λ1

,

λ2

: these roots are known as the characteristic roots or eigenvalues of the characteristic equation. Different solutions are obtained depending on the nature of the roots: If these roots are distinct, we have the general solution

a_n = C\lambda_1^n+D\lambda_2^n

while if they are identical (when

A2+4B=0

), we have

a_n = C\lambda^n + D n\lambda^n

This is the most general solution; the two constants

C

and

D

can be chosen based on two given initial conditions

a0

and

a1

to produce a specific solution.

In the case of complex eigenvalues (which also gives rise to complex values for the solution parameters

C

and

D

), the use of complex numbers can be eliminated by rewriting the solution in trigonometric form. In this case we can write the eigenvalues as

λ1,λ2=\alpha\pm\betai.

Then it can be shown that

a_n = C\lambda_1^n + D\lambda_2^n

can be rewritten as[4]

a_n = 2 M^n \left(E \cos(\theta n) + F \sin(\theta n)\right) = 2 G M^n \cos(\theta n - \delta),

where

\begin M = \sqrt & \cos (\theta) =\tfrac & \sin(\theta) = \tfrac \\ C,D = E \mp F i & & \\ G = \sqrt & \cos (\delta) = \tfrac & \sin (\delta)= \tfrac\end

Here

E

and

F

(or equivalently,

G

and

\delta

) are real constants which depend on the initial conditions. Using \lambda_1+\lambda_2=2 \alpha = A,\lambda_1 \cdot \lambda_2=\alpha^2+\beta^2=-B,

one may simplify the solution given above as

a_n = (-B)^ \left(E \cos(\theta n) + F \sin(\theta n)\right),

where

a1

and

a2

are the initial conditions and

\begin E &= \frac \\ F &=-i \frac \\\theta &=\arccos \left (\frac \right)\end

In this way there is no need to solve for

λ1

and

λ2

.

In all cases—real distinct eigenvalues, real duplicated eigenvalues, and complex conjugate eigenvalues—the equation is stable (that is, the variable

a

converges to a fixed value [specifically, zero]) if and only if both eigenvalues are smaller than one in absolute value. In this second-order case, this condition on the eigenvalues can be shown[5] to be equivalent to

|A|<1-B<2

, which is equivalent to

|B|<1

and

|A|<1-B

.

General solution

Characteristic polynomial and roots

Solving the homogeneous equation

x_t= a_1x_ + \cdots + a_nx_

involves first solving its characteristic polynomial

\lambda^n = a_1\lambda^ + \cdots + a_\lambda^2+a_ \lambda + a_n

for its characteristic roots . These roots can be solved for algebraically if, but not necessarily otherwise. If the solution is to be used numerically, all the roots of this characteristic equation can be found by numerical methods. However, for use in a theoretical context it may be that the only information required about the roots is whether any of them are greater than or equal to 1 in absolute value.

It may be that all the roots are real or instead there may be some that are complex numbers. In the latter case, all the complex roots come in complex conjugate pairs.

Solution with distinct characteristic roots

If all the characteristic roots are distinct, the solution of the homogeneous linear recurrence

x_t= a_1x_ + \cdots + a_nx_

can be written in terms of the characteristic roots as

x_t=c_1\lambda_1^t +\cdots + c_n\lambda_n^t

where the coefficients can be found by invoking the initial conditions. Specifically, for each time period for which an iterate value is known, this value and its corresponding value of can be substituted into the solution equation to obtain a linear equation in the as-yet-unknown parameters; such equations, one for each initial condition, can be solved simultaneously for the parameter values. If all characteristic roots are real, then all the coefficient values will also be real; but with non-real complex roots, in general some of these coefficients will also be non-real.

Converting complex solution to trigonometric form

If there are complex roots, they come in conjugate pairs and so do the complex terms in the solution equation. If two of these complex terms are and, the roots can be written as

\lambda_j, \lambda_=\alpha \pm \beta i =M\left(\frac \pm \fraci\right)

where is the imaginary unit and is the modulus of the roots:

M = \sqrt.

Then the two complex terms in the solution equation can be written as

\beginc_j\lambda_j^t + c_\lambda_^t& = M^t\left(c_j\left(\frac + \fraci\right)^t + c_\left(\frac - \fraci\right)^t\right) \\[6pt]& = M^t\left(c_j\left(\cos\theta + i\sin\theta\right)^t + c_\left(\cos \theta - i\sin\theta\right)^t\right) \\[6pt]& = M^t\bigl(c_j\left(\cos\theta t + i\sin \theta t\right) + c_\left(\cos\theta t - i\sin\theta t\right) \bigr)\end

where is the angle whose cosine is and whose sine is ; the last equality here made use of de Moivre's formula.

Now the process of finding the coefficients and guarantees that they are also complex conjugates, which can be written as . Using this in the last equation gives this expression for the two complex terms in the solution equation:

2M^t\left(\gamma \cos\theta t - \delta \sin\theta t\right)

which can also be written as

2\sqrtM^t \cos(\theta t + \psi)

where is the angle whose cosine is and whose sine is .

Cyclicity

Depending on the initial conditions, even with all roots real the iterates can experience a transitory tendency to go above and below the steady state value. But true cyclicity involves a permanent tendency to fluctuate, and this occurs if there is at least one pair of complex conjugate characteristic roots. This can be seen in the trigonometric form of their contribution to the solution equation, involving and .

Solution with duplicate characteristic roots

In the second-order case, if the two roots are identical, they can both be denoted as and a solution may be of the form

x_t = c_1 \lambda^t + c_2 t \lambda^t.

Solution by conversion to matrix form

An alternative solution method involves converting the th order difference equation to a first-order matrix difference equation. This is accomplished by writing,,, and so on. Then the original single th-order equation

y_t = a_1y_ + a_2y_ + \cdots + a_ny_ + b

can be replaced by the following first-order equations:

\beginw_ & = a_1w_ + a_2w_+\cdots + a_nw_ + b \\w_ & = w_ \\& \,\,\,\vdots \\w_ & =w_.\end

Defining the vector as

\mathbf_i = \beginw_ \\ w_ \\ \vdots \\ w_ \end

this can be put in matrix form as

\mathbf_t = \mathbf\mathbf_+\mathbf

Here is an matrix in which the first row contains and all other rows have a single 1 with all other elements being 0, and is a column vector with first element and with the rest of its elements being 0.

This matrix equation can be solved using the methods in the article Matrix difference equation.In the homogeneous case is a para-permanent of a lower triangular matrix [6]

Solution using generating functions

The recurrence

y_t = a_1y_ + \cdots + a_ny_ + b,

can be solved using the theory of generating functions. First, we write Y(x) = \sum_ y_t x^t. The recurrence is then equivalent to the following generating function equation:

Y(x) = a_1xY(x) + a_2x^2Y(x) + \cdots + a_nx^nY(x) + \frac + p(x)

where

p(x)

is a polynomial of degree at most

n-1

correcting the initial terms.From this equation we can solve to get

Y(x) = \left(\frac + p(x)\right) \cdot \frac.

In other words, not worrying about the exact coefficients,

Y(x)

can be expressed as a rational function

Y(x)=

f(x)
g(x)

.

The closed form can then be derived via partial fraction decomposition. Specifically, if the generating function is written as\frac = \sum_i \frac

then the polynomial

p(x)

determines the initial set of corrections

z(n)

, the denominator

(x-

m
r
i)
determines the exponential term
n
r
i
, and the degree

m

together with the numerator

fi(x)

determine the polynomial coefficient

ki(n)

.

Relation to solution to differential equations

The method for solving linear differential equations is similar to the method above—the "intelligent guess" (ansatz) for linear differential equations with constant coefficients is

eλ

where

λ

is a complex number that is determined by substituting the guess into the differential equation.

This is not a coincidence. Considering the Taylor series of the solution to a linear differential equation:

\sum_^\infin \frac (x-a)^n

it can be seen that the coefficients of the series are given by the

n

-th derivative of

f(x)

evaluated at the point

a

. The differential equation provides a linear difference equation relating these coefficients.

This equivalence can be used to quickly solve for the recurrence relationship for the coefficients in the power series solution of a linear differential equation.

The rule of thumb (for equations in which the polynomial multiplying the first term is non-zero at zero) is that:

y^ \to f[n+k]and more generallyx^m*y^ \to n(n-1)...(n-m+1)f[n+k-m]

Example: The recurrence relationship for the Taylor series coefficients of the equation:

(x^2 + 3x -4)y^ -(3x+1)y^ + 2y = 0

is given by

n(n-1)f[n+1] + 3nf[n+2] -4f[n+3] -3nf[n+1] -f[n+2]+ 2f[n] = 0

or

-4f[n+3] +2nf[n+2] + n(n-4)f[n+1] +2f[n] = 0.

This example shows how problems generally solved using the power series solution method taught in normal differential equation classes can be solved in a much easier way.

Example: The differential equation

ay + by' +cy = 0

has solution

y=e^.

The conversion of the differential equation to a difference equation of the Taylor coefficients is

af[n + 2] + bf[n + 1] + cf[n] = 0.

It is easy to see that the

n

-th derivative of

eax

evaluated at

0

is

an

.

Solving with z-transforms

Certain difference equations - in particular, linear constant coefficient difference equations - can be solved using z-transforms. The z-transforms are a class of integral transforms that lead to more convenient algebraic manipulations and more straightforward solutions. There are cases in which obtaining a direct solution would be all but impossible, yet solving the problem via a thoughtfully chosen integral transform is straightforward.

Stability

In the solution equation

x_t = c_1\lambda_1^t +\cdots + c_n\lambda_n^t,

a term with real characteristic roots converges to 0 as grows indefinitely large if the absolute value of the characteristic root is less than 1. If the absolute value equals 1, the term will stay constant as grows if the root is +1 but will fluctuate between two values if the root is −1. If the absolute value of the root is greater than 1 the term will become larger and larger over time. A pair of terms with complex conjugate characteristic roots will converge to 0 with dampening fluctuations if the absolute value of the modulus of the roots is less than 1; if the modulus equals 1 then constant amplitude fluctuations in the combined terms will persist; and if the modulus is greater than 1, the combined terms will show fluctuations of ever-increasing magnitude.

Thus the evolving variable will converge to 0 if all of the characteristic roots have magnitude less than 1.

If the largest root has absolute value 1, neither convergence to 0 nor divergence to infinity will occur. If all roots with magnitude 1 are real and positive, will converge to the sum of their constant terms ; unlike in the stable case, this converged value depends on the initial conditions; different starting points lead to different points in the long run. If any root is −1, its term will contribute permanent fluctuations between two values. If any of the unit-magnitude roots are complex then constant-amplitude fluctuations of will persist.

Finally, if any characteristic root has magnitude greater than 1, then will diverge to infinity as time goes to infinity, or will fluctuate between increasingly large positive and negative values.

A theorem of Issai Schur states that all roots have magnitude less than 1 (the stable case) if and only if a particular string of determinants are all positive.[2]

If a non-homogeneous linear difference equation has been converted to homogeneous form which has been analyzed as above, then the stability and cyclicality properties of the original non-homogeneous equation will be the same as those of the derived homogeneous form, with convergence in the stable case being to the steady-state value instead of to 0.

See also

Notes and References

  1. Book: Chiang, Alpha . Fundamental Methods of Mathematical Economics . registration . New York . McGraw-Hill . Third . 1984 . 0-07-010813-7 .
  2. Book: Baumol, William . William Baumol

    . William Baumol . Economic Dynamics . registration . New York . Macmillan . Third . 1970 . 0-02-306660-1 .

  3. .
  4. Chiang, Alpha C., Fundamental Methods of Mathematical Economics, third edition, McGraw-Hill, 1984.
  5. Papanicolaou, Vassilis, "On the asymptotic stability of a class of linear difference equations," Mathematics Magazine 69(1), February 1996, 34 - 43.
  6. J. Int. Seq.. Roman. Zatorsky. Taras. Goy. 2016. 19. 16.2.2. Parapermanent of triangular matrices and some general theorems on number sequences .