In probability theory, random element is a generalization of the concept of random variable to more complicated spaces than the simple real line. The concept was introduced by Maurice Fréchet (1948) who commented that the “development of probability theory and expansion of area of its applications have led to necessity to pass from schemes where (random) outcomes of experiments can be described by number or a finite set of numbers, to schemes where outcomes of experiments represent, for example, vectors, functions, processes, fields, series, transformations, and also sets or collections of sets.”[1]
The modern-day usage of “random element” frequently assumes the space of values is a topological vector space, often a Banach or Hilbert space with a specified natural sigma algebra of subsets.[2]
Definition
Let be a probability space, and a measurable space. A random element with values in E is a function X: Ω→E which is -measurable. That is, a function X such that for any , the preimage of B lies in .
Sometimes random elements with values in are called -valued random variables.
Note if , where are the real numbers, and is its Borel σ-algebra, then the definition of random element is the classical definition of random variable.
The definition of a random element with values in a Banach space is typically understood to utilize the smallest -algebra on B for which every bounded linear functional is measurable. An equivalent definition, in this case, to the above, is that a map , from a probability space, is a random element if is a random variable for every bounded linear functional f, or, equivalently, that is weakly measurable.
Examples of random elements
Random variable
A random variable is the simplest type of random element. It is a map is a measurable function from the set of possible outcomes to .
As a real-valued function, often describes some numerical quantity of a given event. E.g. the number of heads after a certain number of coin flips; the heights of different people.
When the image (or range) of is finite or countably infinite, the random variable is called a discrete random variable[3] and its distribution can be described by a probability mass function which assigns a probability to each value in the image of . If the image is uncountably infinite then is called a continuous random variable. In the special case that it is absolutely continuous, its distribution can be described by a probability density function, which assigns probabilities to intervals; in particular, each individual point must necessarily have probability zero for an absolutely continuous random variable. Not all continuous random variables are absolutely continuous,[4] for example a mixture distribution. Such random variables cannot be described by a probability density or a probability mass function.
Random vector
A random vector is a column vector (or its transpose, which is a row vector) whose components are scalar-valued random variables on the same probability space , where is the sample space, is the sigma-algebra (the collection of all events), and is the probability measure (a function returning each event's probability).
Random vectors are often used as the underlying implementation of various types of aggregate random variables, e.g. a random matrix, random tree, random sequence, random process, etc.
Random matrix
A random matrix is a matrix-valued random element. Many important properties of physical systems can be represented mathematically as matrix problems. For example, the thermal conductivity of a lattice can be computed from the dynamical matrix of the particle-particle interactions within the lattice.
Random function
A random function is a type of random element in which a single outcome is selected from some family of functions, where the family consists some class of all maps from the domain to the codomain. For example, the class may be restricted to all continuous functions or to all step functions. The values determined by a random function evaluated at different points from the same realization would not generally be statistically independent but, depending on the model, values determined at the same or different points from different realisations might well be treated as independent.
Random process
A Random process is a collection of random variables, representing the evolution of some system of random values over time. This is the probabilistic counterpart to a deterministic process (or deterministic system). Instead of describing a process which can only evolve in one way (as in the case, for example, of solutions of an ordinary differential equation), in a stochastic or random process there is some indeterminacy: even if the initial condition (or starting point) is known, there are several (often infinitely many) directions in which the process may evolve.
In the simple case of discrete time, as opposed to continuous time, a stochastic process involves a sequence of random variables and the time series associated with these random variables (for example, see Markov chain, also known as discrete-time Markov chain).
Random field
Given a probability space and a measurable space X, an X-valued random field is a collection of X-valued random variables indexed by elements in a topological space T. That is, a random field F is a collection
where each is an X-valued random variable.
Several kinds of random fields exist, among them the Markov random field (MRF), Gibbs random field (GRF), conditional random field (CRF), and Gaussian random field. An MRF exhibits the Markovian property
where is a set of neighbours of the random variable Xi. In other words, the probability that a random variable assumes a value depends on the other random variables only through the ones that are its immediate neighbours. The probability of a random variable in an MRF is given by
where Ω' is the same realization of Ω, except for random variable Xi. It is difficult to calculate with this equation, without recourse to the relation between MRFs and GRFs proposed by Julian Besag in 1974.
Random measure
A random measure is a measure-valued random element.[5][6] Let X be a complete separable metric space and the σ-algebra of its Borel sets. A Borel measure μ on X is boundedly finite if μ(A) < ∞ for every bounded Borel set A. Let be the space of all boundedly finite measures on . Let (Ω, ℱ, P) be a probability space, then a random measure maps from this probability space to the measurable space (, ).[7] A measure generally might be decomposed as:
Here is a diffuse measure without atoms, while is a purely atomic measure.
Random set
A random set is a set-valued random element.
One specific example is a random compact set. Let be a complete separable metric space. Let denote the set of all compact subsets of . The Hausdorff metric on is defined by
is also а complete separable metric space. The corresponding open subsets generate a σ-algebra on , the Borel sigma algebra of .
A random compact set is а measurable function from а probability space into .
Put another way, a random compact set is a measurable function such that is almost surely compact and
is a measurable function for every .
Random geometric objects
These include random points, random figures,[8] and random shapes.[8]
References
- ↑ Fréchet, M. (1948). "Les éléments aléatoires de nature quelconque dans un espace distancié". Annales de l'Institut Henri Poincaré. 10 (4): 215–310.
- ↑ V.V. Buldygin, A.B. Kharazishvili. Geometric Aspects of Probability Theory and Mathematical Statistics. – Kluwer Academic Publishers, Dordrecht. – 2000
- ↑ Yates, Daniel S.; Moore, David S; Starnes, Daren S. (2003). The Practice of Statistics (2nd ed.). New York: Freeman. ISBN 978-0-7167-4773-4. Archived from the original on 2005-02-09.
- ↑ L. Castañeda; V. Arunachalam & S. Dharmaraja (2012). Introduction to Probability and Stochastic Processes with Applications. Wiley. p. 67. ISBN 9781118344941.
- ↑ Kallenberg, O., Random Measures, 4th edition. Academic Press, New York, London; Akademie-Verlag, Berlin (1986). ISBN 0-12-394960-2 MR854102. An authoritative but rather difficult reference.
- ↑ Jan Grandell, Point processes and random measures, Advances in Applied Probability 9 (1977) 502-526. MR0478331 JSTOR A nice and clear introduction.
- ↑ Daley, D. J.; Vere-Jones, D. (2003). An Introduction to the Theory of Point Processes. Probability and its Applications. doi:10.1007/b97277. ISBN 0-387-95541-0.
- 1 2 Stoyan, D., and Stoyan, H. (1994) Fractals, Random Shapes and Point Fields. Methods of Geometrical Statistics. Chichester, New York: John Wiley & Sons. ISBN 0-471-93757-6
Literature
- Hoffman-Jorgensen J., Pisier G. (1976) "Ann.Probab.", v.4, 587–589.
- Mourier E. (1955) Elements aleatoires dans un espace de Banach (These). Paris.
- Prokhorov Yu.V. (1999) Random element. Probability and Mathematical statistics. Encyclopedia. Moscow: "Great Russian Encyclopedia", P.623.