Separation principle explained

In control theory, a separation principle, more formally known as a principle of separation of estimation and control, states that under some assumptions the problem of designing an optimal feedback controller for a stochastic system can be solved by designing an optimal observer for the state of the system, which feeds into an optimal deterministic controller for the system. Thus the problem can be broken into two separate parts, which facilitates the design.

The first instance of such a principle is in the setting of deterministic linear systems, namely that if a stable observer and a stable state feedback are designed for a linear time-invariant system (LTI system hereafter), then the combined observer and feedback is stable. The separation principle does not hold in general for nonlinear systems.

Another instance of the separation principle arises in the setting of linear stochastic systems, namely that state estimation (possibly nonlinear) together with an optimal state feedback controller designed to minimize a quadratic cost, is optimal for the stochastic control problem with output measurements. When process and observation noise are Gaussian, the optimal solution separates into a Kalman filter and a linear-quadratic regulator. This is known as linear-quadratic-Gaussian control. More generally, under suitable conditions and when the noise is a martingale (with possible jumps), again a separation principle applies and is known as the separation principle in stochastic control.[1] [2] [3] [4] [5] [6]

The separation principle also holds for high gain observers used for state estimation of a class of nonlinear systems[7] and control of quantum systems.

Proof of separation principle for deterministic LTI systems

Consider a deterministic LTI system:

\begin{align} x

(t)&=Ax(t)+Bu(t)\\ y(t)&=Cx(t) \end{align}

where

u(t)

represents the input signal,

y(t)

represents the output signal, and

x(t)

represents the internal state of the system.

We can design an observer of the form

\hat{x
} = (A - L C) \hat + B u + L y \,

and state feedback

u(t)=-K\hat{x}.

Define the error e:

e=x-\hat{x}.

Then

e

=(A-LC)e

u(t)=-K(x-e).

Now we can write the closed-loop dynamics as

\begin{bmatrix} x\\
e

\\ \end{bmatrix}=\begin{bmatrix} A-BK&BK\\ 0&A-LC\\ \end{bmatrix} \begin{bmatrix} x\\ e\\ \end{bmatrix}.

Since this is a triangular matrix, the eigenvalues are just those of A - BK together with those of A - LC.[8] Thus the stability of the observer and feedback are independent.

References

Notes and References

  1. Book: Karl Johan Astrom . Introduction to Stochastic Control Theory . Academic Press . 58 . 1970 . 0-486-44531-3.
  2. Tyrone Duncan and Pravin Varaiya . On the solutions of a stochastic control system . SIAM J. Control . 9 . 3 . 354–371 . 1971. 10.1137/0309026 . 1808/16692 . free .
  3. M.H.A. Davis and P. Varaiya . Information states for stochastic systems . J. Math. Anal. Applications . 37 . 384–402 . 1972. 10.1016/0022-247X(72)90281-8 . free .
  4. Anders Lindquist. On Feedback Control of Linear Stochastic Systems . SIAM Journal on Control . 11 . 323–343 . 1973. 2 . 10.1137/0311025 .
  5. Book: A. Bensoussan . Stochastic Control of Partially Observable Systems . Cambridge University Press . 1992.
  6. Tryphon T. Georgiou and Anders Lindquist . The Separation Principle in Stochastic Control, Redux . IEEE Transactions on Automatic Control . 58 . 10 . 2481–2494 . 2013 . 10.1109/TAC.2013.2259207. 1103.3005 . 12623187 .
  7. Book: Atassi. A.N.. Khalil. H.K.. Proceedings of the 37th IEEE Conference on Decision and Control (Cat. No.98CH36171) . A separation principle for the control of a class of nonlinear systems . http://dx.doi.org/10.1109/cdc.1998.760800. 1998. 1. 855–860. IEEE. 10.1109/cdc.1998.760800. 0-7803-4394-8. 126270534.
  8. Proof can be found in this math.stackexchange https://math.stackexchange.com/q/21454.