Random element explained

In probability theory, random element is a generalization of the concept of random variable to more complicated spaces than the simple real line. The concept was introduced by who commented that the “development of probability theory and expansion of area of its applications have led to necessity to pass from schemes where (random) outcomes of experiments can be described by number or a finite set of numbers, to schemes where outcomes of experiments represent, for example, vectors, functions, processes, fields, series, transformations, and also sets or collections of sets.”[1]

The modern-day usage of “random element” frequently assumes the space of values is a topological vector space, often a Banach or Hilbert space with a specified natural sigma algebra of subsets.[2]

Definition

Let

(\Omega,l{F},P)

be a probability space, and

(E,l{E})

a measurable space. A random element with values in E is a function which is

(l{F},l{E})

-measurable. That is, a function X such that for any

B\inl{E}

, the preimage of B lies in

l{F}

.

Sometimes random elements with values in

E

are called

E

-valued random variables.

Note if

(E,l{E})=(R,l{B}(R))

, where

R

are the real numbers, and

l{B}(R)

is its Borel σ-algebra, then the definition of random element is the classical definition of random variable.

The definition of a random element

X

with values in a Banach space

B

is typically understood to utilize the smallest

\sigma

-algebra on B for which every bounded linear functional is measurable. An equivalent definition, in this case, to the above, is that a map

X:\OmegaB

, from a probability space, is a random element if

f\circX

is a random variable for every bounded linear functional f, or, equivalently, that

X

is weakly measurable.

Examples of random elements

Random variable

See main article: Random variable. A random variable is the simplest type of random element. It is a map

X\colon\Omega\toR

is a measurable function from the set of possible outcomes

\Omega

to

R

.

As a real-valued function,

X

often describes some numerical quantity of a given event. E.g. the number of heads after a certain number of coin flips; the heights of different people.

When the image (or range) of

X

is finite or countably infinite, the random variable is called a discrete random variable[3] and its distribution can be described by a probability mass function which assigns a probability to each value in the image of

X

. If the image is uncountably infinite then

X

is called a continuous random variable. In the special case that it is absolutely continuous, its distribution can be described by a probability density function, which assigns probabilities to intervals; in particular, each individual point must necessarily have probability zero for an absolutely continuous random variable. Not all continuous random variables are absolutely continuous,[4] for example a mixture distribution. Such random variables cannot be described by a probability density or a probability mass function.

Random vector

X=(X1,...,X

T
n)
(or its transpose, which is a row vector) whose components are scalar-valued random variables on the same probability space

(\Omega,l{F},P)

, where

\Omega

is the sample space,

l{F}

is the sigma-algebra (the collection of all events), and

P

is the probability measure (a function returning each event's probability).

Random vectors are often used as the underlying implementation of various types of aggregate random variables, e.g. a random matrix, random tree, random sequence, random process, etc.

Random matrix

See main article: Random matrix theory. A random matrix is a matrix-valued random element. Many important properties of physical systems can be represented mathematically as matrix problems. For example, the thermal conductivity of a lattice can be computed from the dynamical matrix of the particle-particle interactions within the lattice.

Random function

See main article: Random function. A random function is a type of random element in which a single outcome is selected from some family of functions, where the family consists some class of all maps from the domain to the codomain. For example, the class may be restricted to all continuous functions or to all step functions. The values determined by a random function evaluated at different points from the same realization would not generally be statistically independent but, depending on the model, values determined at the same or different points from different realisations might well be treated as independent.

Random process

See main article: Random process. A Random process is a collection of random variables, representing the evolution of some system of random values over time. This is the probabilistic counterpart to a deterministic process (or deterministic system). Instead of describing a process which can only evolve in one way (as in the case, for example, of solutions of an ordinary differential equation), in a stochastic or random process there is some indeterminacy: even if the initial condition (or starting point) is known, there are several (often infinitely many) directions in which the process may evolve.

In the simple case of discrete time, as opposed to continuous time, a stochastic process involves a sequence of random variables and the time series associated with these random variables (for example, see Markov chain, also known as discrete-time Markov chain).

Random field

(\Omega,l{F},P)

and a measurable space X,an X-valued random field is a collection of X-valuedrandom variables indexed by elements in a topological space T. That is, a random field F is a collection

\{Ft:t\inT\}

where each

Ft

is an X-valued random variable.

Several kinds of random fields exist, among them the Markov random field (MRF), Gibbs random field (GRF), conditional random field (CRF), and Gaussian random field. An MRF exhibits the Markovian property

P(Xi=xi|Xj=xj,ij)=P(Xi=xi|\partiali),

where

\partiali

is a set of neighbours of the random variable Xi. In other words, the probability that a random variable assumes a value depends on the other random variables only through the ones that are its immediate neighbours. The probability of a random variable in an MRF is given by

P(Xi=xi|\partiali)=

P(\omega)
\sum\omega'P(\omega')

,

where Ω' is the same realization of Ω, except for random variable Xi. It is difficult to calculate with this equation, without recourse to the relation between MRFs and GRFs proposed by Julian Besag in 1974.

Random measure

See main article: Random measure. A random measure is a measure-valued random element.[5] [6] Let X be a complete separable metric space and

ak{B}(X)

the σ-algebra of its Borel sets. A Borel measure μ on X is boundedly finite if μ(A) < ∞ for every bounded Borel set A. Let

MX

be the space of all boundedly finite measures on

ak{B}(X)

. Let be a probability space, then a random measure maps from this probability space to the measurable space .[7] A measure generally might be decomposed as:

\mu=\mud+\mua=\mud+

N
\sum
n=1

\kappan

\delta
Xn

,

Here

\mud

is a diffuse measure without atoms, while

\mua

is a purely atomic measure.

Random set

A random set is a set-valued random element.

One specific example is a random compact set. Let

(M,d)

be a complete separable metric space. Let

l{K}

denote the set of all compact subsets of

M

. The Hausdorff metric

h

on

l{K}

is defined by

h(K1,K2):=max\left\{

\sup
a\inK1
inf
b\inK2

d(a,b),

\sup
b\inK2
inf
a\inK1

d(a,b)\right\}.

(l{K},h)

is also а complete separable metric space. The corresponding open subsets generate a σ-algebra on

l{K}

, the Borel sigma algebra

l{B}(l{K})

of

l{K}

.

K

from а probability space

(\Omega,l{F},P)

into

(l{K},l{B}(l{K}))

.

Put another way, a random compact set is a measurable function

K\colon\Omega\to2M

such that

K(\omega)

is almost surely compact and

\omega\mapstoinfbd(x,b)

is a measurable function for every

x\inM

.

Random geometric objects

These include random points, random figures,[8] and random shapes.[8]

References

Literature

External links

Notes and References

  1. Fréchet . M. . Maurice Fréchet . 1948 . Les éléments aléatoires de nature quelconque dans un espace distancié . . 10 . 4 . 215–310 .
  2. V.V. Buldygin, A.B. Kharazishvili. Geometric Aspects of Probability Theory and Mathematical Statistics. – Kluwer Academic Publishers, Dordrecht. – 2000
  3. Book: Yates . Daniel S. . Moore . David S . Starnes . Daren S. . 2003 . The Practice of Statistics . 2nd . . New York . 978-0-7167-4773-4 . dead . https://web.archive.org/web/20050209001108/http://bcs.whfreeman.com/yates2e/ . 2005-02-09 .
  4. Book: L. Castañeda . V. Arunachalam . S. Dharmaraja . amp . Introduction to Probability and Stochastic Processes with Applications . 2012 . Wiley . 67 . 9781118344941 .
  5. [Olav Kallenberg|Kallenberg, O.]
  6. Jan Grandell, Point processes and random measures, Advances in Applied Probability 9 (1977) 502-526. JSTOR A nice and clear introduction.
  7. Book: 10.1007/b97277 . D. J. . Daley . D. . Vere-Jones. An Introduction to the Theory of Point Processes . Probability and its Applications . 2003 . 0-387-95541-0 .
  8. Stoyan, D., and Stoyan, H. (1994) Fractals, Random Shapes and Point Fields. Methods of Geometrical Statistics. Chichester, New York: John Wiley & Sons.