Measure-preserving dynamical system
In mathematics, a measure-preserving dynamical system is an object of study in the abstract formulation of dynamical systems, and ergodic theory in particular. Measure-preserving systems obey the Poincaré recurrence theorem, and are a special case of conservative systems. They provide the formal, mathematical basis for a broad range of physical systems, and, in particular, many systems from classical mechanics (in particular, most non-dissipative systems) as well as systems in thermodynamic equilibrium.
Definition
A measure-preserving dynamical system is defined as a probability space and a measure-preserving transformation on it. In more detail, it is a system
- [math]\displaystyle{ (X, \mathcal{B}, \mu, T) }[/math]
with the following structure:
- [math]\displaystyle{ X }[/math] is a set,
- [math]\displaystyle{ \mathcal B }[/math] is a σ-algebra over [math]\displaystyle{ X }[/math],
- [math]\displaystyle{ \mu:\mathcal{B}\rightarrow[0,1] }[/math] is a probability measure, so that [math]\displaystyle{ \mu (X) = 1 }[/math], and [math]\displaystyle{ \mu(\varnothing) = 0 }[/math],
- [math]\displaystyle{ T:X \rightarrow X }[/math] is a measurable transformation which preserves the measure [math]\displaystyle{ \mu }[/math], i.e., [math]\displaystyle{ \forall A\in \mathcal{B}\;\; \mu(T^{-1}(A))=\mu(A) }[/math].
Discussion
One may ask why the measure preserving transformation is defined in terms of the inverse [math]\displaystyle{ \mu(T^{-1}(A))=\mu(A) }[/math] instead of the forward transformation [math]\displaystyle{ \mu(T(A))=\mu(A) }[/math]. This can be understood intuitively.
Consider the typical measure on the unit interval [math]\displaystyle{ [0, 1] }[/math], and a map [math]\displaystyle{ Tx = 2x\mod 1 = \begin{cases} 2x \text{ if } x \lt 1/2 \\ 2x-1 \text{ if } x \gt 1/2 \\ \end{cases} }[/math]. This is the Bernoulli map. Now, distribute an even layer of paint on the unit interval [math]\displaystyle{ [0, 1] }[/math], and then map the paint forward. The paint on the [math]\displaystyle{ [0, 1/2] }[/math] half is spread thinly over all of [math]\displaystyle{ [0, 1] }[/math], and the paint on the [math]\displaystyle{ [1/2, 1] }[/math] half as well. The two layers of thin paint, layered together, recreates the exact same paint thickness.
More generally, the paint that would arrive at subset [math]\displaystyle{ A \subset [0, 1] }[/math] comes from the subset [math]\displaystyle{ T^{-1}(A) }[/math]. For the paint thickness to remain unchanged (measure-preserving), the mass of incoming paint should be the same: [math]\displaystyle{ \mu(A) = \mu(T^{-1}(A)) }[/math].
Consider a mapping [math]\displaystyle{ \mathcal{T} }[/math] of power sets:
- [math]\displaystyle{ \mathcal{T}:P(X)\to P(X) }[/math]
Consider now the special case of maps [math]\displaystyle{ \mathcal{T} }[/math] which preserve intersections, unions and complements (so that it is a map of Borel sets) and also sends [math]\displaystyle{ X }[/math] to [math]\displaystyle{ X }[/math] (because we want it to be conservative). Every such conservative, Borel-preserving map can be specified by some surjective map [math]\displaystyle{ T:X\to X }[/math] by writing [math]\displaystyle{ \mathcal{T}(A)=T^{-1}(A) }[/math]. Of course, one could also define [math]\displaystyle{ \mathcal{T}(A)=T(A) }[/math], but this is not enough to specify all such possible maps [math]\displaystyle{ \mathcal{T} }[/math]. That is, conservative, Borel-preserving maps [math]\displaystyle{ \mathcal{T} }[/math] cannot, in general, be written in the form [math]\displaystyle{ \mathcal{T}(A)=T(A); }[/math].
[math]\displaystyle{ \mu(T^{-1}(A)) }[/math] has the form of a pushforward, whereas [math]\displaystyle{ \mu(T(A)) }[/math] is generically called a pullback. Almost all properties and behaviors of dynamical systems are defined in terms of the pushforward. For example, the transfer operator is defined in terms of the pushforward of the transformation map [math]\displaystyle{ T }[/math]; the measure [math]\displaystyle{ \mu }[/math] can now be understood as an invariant measure; it is just the Frobenius–Perron eigenvector of the transfer operator (recall, the FP eigenvector is the largest eigenvector of a matrix; in this case it is the eigenvector which has the eigenvalue one: the invariant measure.)
There are two classification problems of interest. One, discussed below, fixes [math]\displaystyle{ (X, \mathcal{B}, \mu) }[/math] and asks about the isomorphism classes of a transformation map [math]\displaystyle{ T }[/math]. The other, discussed in transfer operator, fixes [math]\displaystyle{ (X, \mathcal{B}) }[/math] and [math]\displaystyle{ T }[/math], and asks about maps [math]\displaystyle{ \mu }[/math] that are measure-like. Measure-like, in that they preserve the Borel properties, but are no longer invariant; they are in general dissipative and so give insights into dissipative systems and the route to equilibrium.
In terms of physics, the measure-preserving dynamical system [math]\displaystyle{ (X, \mathcal{B}, \mu, T) }[/math] often describes a physical system that is in equilibrium, for example, thermodynamic equilibrium. One might ask: how did it get that way? Often, the answer is by stirring, mixing, turbulence, thermalization or other such processes. If a transformation map [math]\displaystyle{ T }[/math] describes this stirring, mixing, etc. then the system [math]\displaystyle{ (X, \mathcal{B}, \mu, T) }[/math] is all that is left, after all of the transient modes have decayed away. The transient modes are precisely those eigenvectors of the transfer operator that have eigenvalue less than one; the invariant measure [math]\displaystyle{ \mu }[/math] is the one mode that does not decay away. The rate of decay of the transient modes are given by (the logarithm of) their eigenvalues; the eigenvalue one corresponds to infinite half-life.
Informal example
The microcanonical ensemble from physics provides an informal example. Consider, for example, a fluid, gas or plasma in a box of width, length and height [math]\displaystyle{ w\times l\times h, }[/math] consisting of [math]\displaystyle{ N }[/math] atoms. A single atom in that box might be anywhere, having arbitrary velocity; it would be represented by a single point in [math]\displaystyle{ w\times l\times h\times \mathbb{R}^3. }[/math] A given collection of [math]\displaystyle{ N }[/math] atoms would then be a single point somewhere in the space [math]\displaystyle{ (w\times l\times h)^N \times \mathbb{R}^{3N}. }[/math] The "ensemble" is the collection of all such points, that is, the collection of all such possible boxes (of which there are an uncountably-infinite number). This ensemble of all-possible-boxes is the space [math]\displaystyle{ X }[/math] above.
In the case of an ideal gas, the measure [math]\displaystyle{ \mu }[/math] is given by the Maxwell–Boltzmann distribution. It is a product measure, in that if [math]\displaystyle{ p_i(x,y,z,v_x,v_y,v_z)\,d^3x\,d^3p }[/math] is the probability of atom [math]\displaystyle{ i }[/math] having position and velocity [math]\displaystyle{ x,y,z,v_x,v_y,v_z }[/math], then, for [math]\displaystyle{ N }[/math] atoms, the probability is the product of [math]\displaystyle{ N }[/math] of these. This measure is understood to apply to the ensemble. So, for example, one of the possible boxes in the ensemble has all of the atoms on one side of the box. One can compute the likelihood of this, in the Maxwell–Boltzmann measure. It will be enormously tiny, of order [math]\displaystyle{ \mathcal{O}\left(2^{-3N}\right). }[/math] Of all possible boxes in the ensemble, this is a ridiculously small fraction.
The only reason that this is an "informal example" is because writing down the transition function [math]\displaystyle{ T }[/math] is difficult, and, even if written down, it is hard to perform practical computations with it. Difficulties are compounded if the interaction is not an ideal-gas billiard-ball type interaction, but is instead a van der Waals interaction, or some other interaction suitable for a liquid or a plasma; in such cases, the invariant measure is no longer the Maxwell–Boltzmann distribution. The art of physics is finding reasonable approximations.
This system does exhibit one key idea from the classification of measure-preserving dynamical systems: two ensembles, having different temperatures, are inequivalent. The entropy for a given canonical ensemble depends on its temperature; as physical systems, it is "obvious" that when the temperatures differ, so do the systems. This holds in general: systems with different entropy are not isomorphic.
Examples
Unlike the informal example above, the examples below are sufficiently well-defined and tractable that explicit, formal computations can be performed.
- μ could be the normalized angle measure dθ/2π on the unit circle, and T a rotation. See equidistribution theorem;
- the Bernoulli scheme;
- the interval exchange transformation;
- with the definition of an appropriate measure, a subshift of finite type;
- the base flow of a random dynamical system;
- the flow of a Hamiltonian vector field on the tangent bundle of a closed connected smooth manifold is measure-preserving (using the measure induced on the Borel sets by the symplectic volume form) by Liouville's theorem (Hamiltonian);[1]
- for certain maps and Markov processes, the Krylov–Bogolyubov theorem establishes the existence of a suitable measure to form a measure-preserving dynamical system.
Generalization to groups and monoids
The definition of a measure-preserving dynamical system can be generalized to the case in which T is not a single transformation that is iterated to give the dynamics of the system, but instead is a monoid (or even a group, in which case we have the action of a group upon the given probability space) of transformations Ts : X → X parametrized by s ∈ Z (or R, or N ∪ {0}, or [0, +∞)), where each transformation Ts satisfies the same requirements as T above.[1] In particular, the transformations obey the rules:
- [math]\displaystyle{ T_0 = \mathrm{id}_X :X \rightarrow X }[/math], the identity function on X;
- [math]\displaystyle{ T_{s} \circ T_{t} = T_{t + s} }[/math], whenever all the terms are well-defined;
- [math]\displaystyle{ T_{s}^{-1} = T_{-s} }[/math], whenever all the terms are well-defined.
The earlier, simpler case fits into this framework by defining Ts = Ts for s ∈ N.
Homomorphisms
The concept of a homomorphism and an isomorphism may be defined.
Consider two dynamical systems [math]\displaystyle{ (X, \mathcal{A}, \mu, T) }[/math] and [math]\displaystyle{ (Y, \mathcal{B}, \nu, S) }[/math]. Then a mapping
- [math]\displaystyle{ \varphi:X \to Y }[/math]
is a homomorphism of dynamical systems if it satisfies the following three properties:
- The map [math]\displaystyle{ \varphi\ }[/math] is measurable.
- For each [math]\displaystyle{ B \in \mathcal{B} }[/math], one has [math]\displaystyle{ \mu (\varphi^{-1}B) = \nu(B) }[/math].
- For [math]\displaystyle{ \mu }[/math]-almost all [math]\displaystyle{ x \in X }[/math], one has [math]\displaystyle{ \varphi(Tx) = S(\varphi x) }[/math].
The system [math]\displaystyle{ (Y, \mathcal{B}, \nu, S) }[/math] is then called a factor of [math]\displaystyle{ (X, \mathcal{A}, \mu, T) }[/math].
The map [math]\displaystyle{ \varphi\; }[/math] is an isomorphism of dynamical systems if, in addition, there exists another mapping
- [math]\displaystyle{ \psi:Y \to X }[/math]
that is also a homomorphism, which satisfies
- for [math]\displaystyle{ \mu }[/math]-almost all [math]\displaystyle{ x \in X }[/math], one has [math]\displaystyle{ x = \psi(\varphi x) }[/math];
- for [math]\displaystyle{ \nu }[/math]-almost all [math]\displaystyle{ y \in Y }[/math], one has [math]\displaystyle{ y = \varphi(\psi y) }[/math].
Hence, one may form a category of dynamical systems and their homomorphisms.
Generic points
A point x ∈ X is called a generic point if the orbit of the point is distributed uniformly according to the measure.
Symbolic names and generators
Consider a dynamical system [math]\displaystyle{ (X, \mathcal{B}, T, \mu) }[/math], and let Q = {Q1, ..., Qk} be a partition of X into k measurable pair-wise disjoint sets. Given a point x ∈ X, clearly x belongs to only one of the Qi. Similarly, the iterated point Tnx can belong to only one of the parts as well. The symbolic name of x, with regards to the partition Q, is the sequence of integers {an} such that
- [math]\displaystyle{ T^nx \in Q_{a_n}. }[/math]
The set of symbolic names with respect to a partition is called the symbolic dynamics of the dynamical system. A partition Q is called a generator or generating partition if μ-almost every point x has a unique symbolic name.
Operations on partitions
Given a partition Q = {Q1, ..., Qk} and a dynamical system [math]\displaystyle{ (X, \mathcal{B}, T, \mu) }[/math], define the T-pullback of Q as
- [math]\displaystyle{ T^{-1}Q = \{T^{-1}Q_1,\ldots,T^{-1}Q_k\}. }[/math]
Further, given two partitions Q = {Q1, ..., Qk} and R = {R1, ..., Rm}, define their refinement as
- [math]\displaystyle{ Q \vee R = \{Q_i \cap R_j \mid i=1,\ldots,k,\ j=1,\ldots,m,\ \mu(Q_i \cap R_j) \gt 0 \}. }[/math]
With these two constructs, the refinement of an iterated pullback is defined as
- [math]\displaystyle{ \begin{align} \bigvee_{n=0}^N T^{-n}Q & = \{Q_{i_0} \cap T^{-1}Q_{i_1} \cap \cdots \cap T^{-N}Q_{i_N} \\ & {} \qquad \mbox { where }i_\ell = 1,\ldots,k ,\ \ell=0,\ldots,N,\ \\ & {} \qquad \qquad \mu \left (Q_{i_0} \cap T^{-1}Q_{i_1} \cap \cdots \cap T^{-N}Q_{i_N} \right )\gt 0 \} \\ \end{align} }[/math]
which plays crucial role in the construction of the measure-theoretic entropy of a dynamical system.
Measure-theoretic entropy
The entropy of a partition [math]\displaystyle{ \mathcal{Q} }[/math] is defined as[2][3]
- [math]\displaystyle{ H(\mathcal{Q})=-\sum_{Q \in \mathcal{Q}}\mu (Q) \log \mu(Q). }[/math]
The measure-theoretic entropy of a dynamical system [math]\displaystyle{ (X, \mathcal{B}, T, \mu) }[/math] with respect to a partition Q = {Q1, ..., Qk} is then defined as
- [math]\displaystyle{ h_\mu(T,\mathcal{Q}) = \lim_{N \rightarrow \infty} \frac{1}{N} H\left(\bigvee_{n=0}^N T^{-n}\mathcal{Q}\right). }[/math]
Finally, the Kolmogorov–Sinai metric or measure-theoretic entropy of a dynamical system [math]\displaystyle{ (X, \mathcal{B},T,\mu) }[/math] is defined as
- [math]\displaystyle{ h_\mu(T) = \sup_{\mathcal{Q}} h_\mu(T,\mathcal{Q}). }[/math]
where the supremum is taken over all finite measurable partitions. A theorem of Yakov Sinai in 1959 shows that the supremum is actually obtained on partitions that are generators. Thus, for example, the entropy of the Bernoulli process is log 2, since almost every real number has a unique binary expansion. That is, one may partition the unit interval into the intervals [0, 1/2) and [1/2, 1]. Every real number x is either less than 1/2 or not; and likewise so is the fractional part of 2nx.
If the space X is compact and endowed with a topology, or is a metric space, then the topological entropy may also be defined.
If [math]\displaystyle{ T }[/math] is an ergodic, piecewise expanding, and Markov on [math]\displaystyle{ X \subset \R }[/math], and [math]\displaystyle{ \mu }[/math] is absolutely continuous with respect to the Lebesgue measure, then we have the Rokhlin formula[4] (section 4.3 and section 12.3 [5]):[math]\displaystyle{ h_{\mu }(T) = \int \ln |dT/dx| \mu(dx) }[/math]This allows calculation of entropy of many interval maps, such as the logistic map.
Ergodic means that [math]\displaystyle{ T^{-1}(A) = A }[/math] implies [math]\displaystyle{ A }[/math] has full measure or zero measure. Piecewise expanding and Markov means that there is a partition of [math]\displaystyle{ X }[/math] into finitely many open intervals, such that for some [math]\displaystyle{ \epsilon \gt 0 }[/math], [math]\displaystyle{ |T'| \geq 1 + \epsilon }[/math] on each open interval. Markov means that for each [math]\displaystyle{ I_i }[/math] from those open intervals, either [math]\displaystyle{ T(I_i) \cap I_i = \emptyset }[/math] or [math]\displaystyle{ T(I_i) \cap I_i = I_i }[/math].
Classification and anti-classification theorems
One of the primary activities in the study of measure-preserving systems is their classification according to their properties. That is, let [math]\displaystyle{ (X, \mathcal{B}, \mu) }[/math] be a measure space, and let [math]\displaystyle{ U }[/math] be the set of all measure preserving systems [math]\displaystyle{ (X, \mathcal{B}, \mu, T) }[/math]. An isomorphism [math]\displaystyle{ S\sim T }[/math] of two transformations [math]\displaystyle{ S, T }[/math] defines an equivalence relation [math]\displaystyle{ \mathcal{R}\subset U\times U. }[/math] The goal is then to describe the relation [math]\displaystyle{ \mathcal{R} }[/math]. A number of classification theorems have been obtained; but quite interestingly, a number of anti-classification theorems have been found as well. The anti-classification theorems state that there are more than a countable number of isomorphism classes, and that a countable amount of information is not sufficient to classify isomorphisms.[6][7]
The first anti-classification theorem, due to Hjorth, states that if [math]\displaystyle{ U }[/math] is endowed with the weak topology, then the set [math]\displaystyle{ \mathcal{R} }[/math] is not a Borel set.[8] There are a variety of other anti-classification results. For example, replacing isomorphism with Kakutani equivalence, it can be shown that there are uncountably many non-Kakutani equivalent ergodic measure-preserving transformations of each entropy type.[9]
These stand in contrast to the classification theorems. These include:
- Ergodic measure-preserving transformations with a pure point spectrum have been classified.[10]
- Bernoulli shifts are classified by their metric entropy.[11][12][13] See Ornstein theory for more.
Krieger finite generator theorem[14] (Krieger 1970) — Given a dynamical system on a Lebesgue space of measure 1, where [math]\displaystyle{ T }[/math] is invertible, measure preserving, and ergodic.
If [math]\displaystyle{ h_T \leq \ln k }[/math] for some integer [math]\displaystyle{ k }[/math], then the system has a size-[math]\displaystyle{ k }[/math] generator.
If the entropy is exactly equal to [math]\displaystyle{ \ln k }[/math], then such a generator exists iff the system is isomorphic to the Bernoulli shift on [math]\displaystyle{ k }[/math] symbols with equal measures.
See also
- Krylov–Bogolyubov theorem on the existence of invariant measures
- Poincaré recurrence theorem – Certain dynamical systems will eventually return to (or approximate) their initial state
References
- ↑ 1.0 1.1 Walters, Peter (2000). An Introduction to Ergodic Theory. Springer. ISBN 0-387-95152-0.
- ↑ Sinai, Ya. G. (1959). "On the Notion of Entropy of a Dynamical System". Doklady Akademii Nauk SSSR 124: 768–771.
- ↑ Sinai, Ya. G. (2007). "Metric Entropy of Dynamical System". https://web.math.princeton.edu/facultypapers/Sinai/MetricEntropy2.pdf.
- ↑ The Shannon-McMillan-Breiman Theorem
- ↑ Pollicott, Mark; Yuri, Michiko (1998). Dynamical Systems and Ergodic Theory. London Mathematical Society Student Texts. Cambridge: Cambridge University Press. ISBN 978-0-521-57294-1. https://www.cambridge.org/core/books/dynamical-systems-and-ergodic-theory/3C1AA7BE85F5D2EE027D60CC72FDBEB8.
- ↑ Foreman, Matthew; Weiss, Benjamin (2019). "From Odometers to Circular Systems: A Global Structure Theorem". Journal of Modern Dynamics 15: 345–423. doi:10.3934/jmd.2019024.
- ↑ Foreman, Matthew; Weiss, Benjamin (2022). "Measure preserving Diffeomorphisms of the Torus are unclassifiable". Journal of the European Mathematical Society 24 (8): 2605–2690. doi:10.4171/JEMS/1151.
- ↑ Hjorth, G. (2001). "On invariants for measure preserving transformations". Fund. Math. 169 (1): 51–84. doi:10.4064/FM169-1-2.
- ↑ Ornstein, D.; Rudolph, D.; Weiss, B. (1982). Equivalence of measure preserving transformations. Mem. American Mathematical Soc.. 37. ISBN 0-8218-2262-4.
- ↑ Halmos, P.; von Neumann, J. (1942). "Operator methods in classical mechanics. II.". Annals of Mathematics. (2) 43 (2): 332–350. doi:10.2307/1968872.
- ↑ Sinai, Ya. (1962). "A weak isomorphism of transformations with invariant measure". Doklady Akademii Nauk SSSR 147: 797–800.
- ↑ Ornstein, D. (1970). "Bernoulli shifts with the same entropy are isomorphic". Advances in Mathematics 4 (3): 337–352. doi:10.1016/0001-8708(70)90029-0.
- ↑ Katok, A.; Hasselblatt, B. (1995). "Introduction to the modern theory of dynamical systems". Encyclopedia of Mathematics and its Applications. 54. Cambridge University Press.
- ↑ Downarowicz, Tomasz (2011). Entropy in dynamical systems. New Mathematical Monographs. Cambridge: Cambridge University Press. p. 106. ISBN 978-0-521-88885-1.
Further reading
- Michael S. Keane, "Ergodic theory and subshifts of finite type", (1991), appearing as Chapter 2 in Ergodic Theory, Symbolic Dynamics and Hyperbolic Spaces, Tim Bedford, Michael Keane and Caroline Series, Eds. Oxford University Press, Oxford (1991). ISBN 0-19-853390-X (Provides expository introduction, with exercises, and extensive references.)
- Lai-Sang Young, "Entropy in Dynamical Systems" (pdf; ps), appearing as Chapter 16 in Entropy, Andreas Greven, Gerhard Keller, and Gerald Warnecke, eds. Princeton University Press, Princeton, NJ (2003). ISBN 0-691-11338-6
- T. Schürmann and I. Hoffmann, The entropy of strange billiards inside n-simplexes. J. Phys. A 28(17), page 5033, 1995. PDF-Document (gives a more involved example of measure-preserving dynamical system.)
Original source: https://en.wikipedia.org/wiki/Measure-preserving dynamical system.
Read more |