Namespaces
Variants
Actions

Entropy theory of a dynamical system

From Encyclopedia of Mathematics
Revision as of 19:37, 5 June 2020 by Ulf Rehmann (talk | contribs) (tex encoded by computer)
Jump to: navigation, search


2020 Mathematics Subject Classification: Primary: 37A35 Secondary: 60G10 [MSN][ZBL]

A branch of ergodic theory closely connected with probability theory and information theory. In broad lines, the nature of this connection is as follows.

Let $ \{ T _ {t} \} $ be a dynamical system (usually a measurable flow or a cascade) with phase space $ W $ and invariant measure $ \mu $. Let $ f : W \rightarrow \mathbf R $ be a measurable function and let $ \xi $ be the measurable decomposition (measurable partition) of $ W $ into inverse images $ f ^ { - 1 } ( c) $, $ c \in \mathbf R $. (For what follows it is sufficient to consider inverse images of $ f $ having a countable, and as a rule even finite, number of values, and the corresponding partition $ \xi $.) Then

$$ \{ t \mapsto f ( T _ {t} w ) \} $$

is a stationary stochastic process (in the narrow sense of the word) with $ W $ as space of elementary events. Usually this can be regarded as a process $ \{ X _ {t} ( \omega ) \} $ the space of elementary events of which is the space $ \Omega $ of sample functions (cf. Sample function) $ \omega $, endowed with a suitable measure $ \nu $, and $ X _ {t} ( \omega ) = \omega ( t) $. The mapping

$$ \pi : W \rightarrow \Omega ,\ \ ( \pi w ) ( t) = f ( T _ {t} w ) $$

is a homomorphism of measure spaces (see the definition in the article Metric isomorphism) that carries $ \{ T _ {t} \} $ into the shift $ \{ S _ {t} \} $, where $ ( S _ {t} \omega ) ( \tau ) = \omega ( t + \tau ) $.

The process $ \{ X _ {t} ( \omega ) \} $ contains some information about the original system $ \{ T _ {t} \} $. This can even be complete information when $ \pi $ is an isomorphism. (One says then that $ \xi $ is a generator for $ \{ T _ {t} \} $; if $ T $ is an automorphism, then the partition is called a one-side generator for $ T $ if it is a generator for the cascade $ \{ {T ^ {n} } : {n \geq 0 } \} $, and a two-side generator for $ T $ if it is a generator for $ \{ {T ^ {n} } : {n \in \mathbf Z } \} $.) However, $ \{ X _ {t} ( \omega ) \} $ also depends on the choice of $ f $, that is, first of all, on $ \xi $( the specific values of $ f $ on the elements of $ \xi $ are less important here). Of interest in ergodic theory are those properties of an individual process $ \{ X _ {t} ( \omega ) \} $, or of a collection of those processes (obtained for various $ \xi $), that are properties of the system $ \{ T _ {t} \} $ itself. However, to select such properties was for a long time not easy unless they reduced to known ones.

This difficulty was successfully overcome in the middle of the 1950s by A.N. Kolmogorov when he introduced a fundamentally new (non-spectral) invariant, the metric entropy of a dynamical system, and emphasized the role of increasing measurable partitions $ \eta $, that is, those for which $ T _ {t} \eta $ is finer than $ \eta $( $ \mathop{\rm mod} 0 $) for $ t > 0 $. (In this way a partition describes the "past" of the process $ \{ X _ {t} ( \omega ) \} $, see also $ K $- system; Exact endomorphism.) The elaboration of this range of problems (including that of the existence and properties of generating partitions) is the object of the entropy theory of dynamical systems in the form in which it was put together in the middle of the 1960s (see [R]). A substantial addition was the more complete and somewhat more special theory of D. Ornstein in which auxiliary stochastic processes $ \{ X _ {t} ( \omega ) \} $ are used in a more direct way (see [O]). In view of the need to ensure the invariance under metric isomorphisms in both the "Kolmogorov" and the "Ornstein" entropy theory of dynamical systems, probability and information-theoretical ideas pervade the area in an essentially transformed form.

Two conditions of "regularity" type of a stochastic process occurring in the entropy theory of dynamical systems may serve as examples. One of them leads to the definition of a $ K $- system. The other, more restrictive one, a very weak Bernoulli property, turns out to be necessary and sufficient for a shift in the space of sample functions to be isomorphic to a Bernoulli automorphism. It can be verified in a number of examples, the original definitions of which have no relation to stochastic processes.

References

[R] V.A. Rokhlin, "Lectures on the entropy theory of measure-preserving transformations" Russian Math. Surveys , 22 : 5 (1967) pp. 1–52 Uspekhi Mat. Nauk , 22 : 5 (1967) pp. 3–56 Zbl 0174.45501
[O] D. Ornstein, "Ergodic theory, randomness, and dynamical systems" , Yale Univ. Press (1974) MR0447525 Zbl 0296.28016

See also the references to $ K $- system; Entropy; Ergodic theory.

Comments

Entropy in the theory of dynamical systems is defined as follows (cf. also Entropy). For every measurable partition $ \xi = \{ A _ {1} \dots A _ {m} \} $ of a probability space $ ( W, {\mathcal A} , \mu ) $, the entropy $ H ( \xi ) $ of $ \xi $ is defined as

$$ H ( \xi ) = - \sum _ { k= } 1 ^ { m } \mu ( A _ {k} ) \mathop{\rm log} \mu ( A _ {k} ) $$

(it is assumed that $ 0 \mathop{\rm log} 0 = 0 $). The base of the logarithm can be any positive number, but as a rule one takes logarithms to the base 2 or $ e $.

Then define the entropy $ h ( T, \xi ) $ of a measure-preserving transformation $ T $( i.e. of a cascade) with respect to a partition $ \xi $ by

$$ h ( T, \xi ) = \lim\limits _ {n \rightarrow \infty } \frac{1}{n} H ( \xi ^ {n} ) , $$

where $ \xi ^ {n} = \lor _ {i=} 1 ^ {n-} 1 T ^ { - i } \xi $, the common refinement of the partitions $ \xi , T ^ { - 1 } \xi \dots T ^ { - n } \xi $. Finally, the entropy of $ T $ is defined as

$$ h ( T) = \sup h ( T , \xi ), $$

where the supremum is over all finite measurable partitions of $ W $.

Since for a flow $ \{ T _ {t} \} $ in $ W $ one has $ h ( T _ {t} ) = | t | h ( T _ {1} ) $ for all $ t \in \mathbf R $, one usually defines the entropy of a flow $ \{ T _ {t} \} $ by

$$ h ( \{ T _ {t} \} ) = h ( T _ {1} ). $$

References

[CFS] I.P. Cornfel'd, S.V. Fomin, Ya.G. Sinai, "Ergodic theory" , Springer (1982) (Translated from Russian) MR832433
[M] R. Mañé, "Ergodic theory and differentiable dynamics" , Springer (1987) MR0889254 Zbl 0616.28007
How to Cite This Entry:
Entropy theory of a dynamical system. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Entropy_theory_of_a_dynamical_system&oldid=26441
This article was adapted from an original article by D.V. Anosov (originator), which appeared in Encyclopedia of Mathematics - ISBN 1402006098. See original article