Sequential estimation explained

In statistics, sequential estimation refers to estimation methods in sequential analysis where the sample size is not fixed in advance. Instead, data is evaluated as it is collected, and further sampling is stopped in accordance with a predefined stopping rule as soon as significant results are observed.The generic version is called the optimal Bayesian estimator,[1] which is the theoretical underpinning for every sequential estimator (but cannot be instantiated directly). It includes a Markov process for the state propagation and measurement process for each state, which yields some typical statistical independence relations. The Markov process describes the propagation of a probability distribution over discrete time instances and the measurement is the information one has about each time instant, which is usually less informative than the state. Only the observed sequence will, together with the models, accumulate the information of all measurements and the corresponding Markov process to yield better estimates.

From that, the Kalman filter (and its variants), the particle filter, the histogram filter and others can be derived. It depends on the models, which one to use and requires experience to choose the right one. In most cases, the goal is to estimate the state sequence from the measurements. In other cases, one can use the description to estimate the parameters of a noise process for example. One can also accumulate the unmodeled statistical behavior of the states projected in the measurement space (called innovation sequence, which naturally includes the orthogonality principle in its derivations to yield an independence relation and therefore can be also cast into a Hilbert space representation, which makes it very intuitive) over time and compare it with a threshold, which then corresponds to the aforementioned stopping criterion. One difficulty is to set up the initial conditions for the probabilistic models, which is in most cases done by experience, data sheets or precise measurements with a different setup.

The statistical behaviour of the heuristic/sampling methods (e.g. particle filter or histogram filter) depends on many parameters and implementation details and should not be used in safety critical applications (since it is very hard to yield theoretical guarantees or do proper testing), unless one has a very good reason.

If there is a dependence of each state on an overall entity (e.g. a map or simply an overall state variable), one typically uses SLAM (simultaneous localization and mapping) techniques, which include the sequential estimator as a special case (when the overall state variable has just one state). It will estimate the state sequence and the overall entity.

There are also none-causal variants, that have all measurements at the same time, batches of measurements or revert the state evolution to go backwards again. These are then, however, not real time capable (except one uses a really big buffer, that lowers the throughput dramatically) anymore and only sufficient for post processing. Other variants do several passes to yield a rough estimate first and then refine it by the following passes, which is inspired by video editing/transcoding. For image processing (where all pixels are available at the same time) these methods become causal again.

Sequential estimation is the core of many well known applications, such as the Viterbi decoder, convolutional codes, video compression or target tracking. Due to its state space representation, which is in most cases motivated by physical laws of motion, there is a direct link to control applications, which led to the use of the Kalman filter for space applications for example.

See also

References

. Abraham Wald . Sequential Analysis . 1947 . . . 0-471-91806-7 . See Dover reprint: .

  1. Web site: Bayesian Estimation Lecture Notes .