Entropy dimensions and a class of constructive examples

Entropy dimensions and a class of constructive examples∗
Sébastien Ferenczi
Institut de Mathématiques de Luminy
CNRS - UMR 6206
Case 907, 163 av. de Luminy
F13288 Marseille Cedex 9 (France)
and Fédération de Recherche des Unités de Mathématiques de Marseille
CNRS - FR 2291
[email protected]
Kyewon Koh Park
Department of Mathematics
Ajou University
Suwon 442-729
Korea
[email protected]
December 9, 2004
Abstract
Motivated by the study of actions of ZZ2 and more general groups, and their noncocompact subgroup actions, we investigate entropy-type invariants for deterministic
systems. In particular, we define a new isomorphism invariant, the entropy dimension,
and look at its behaviour on examples. We also look at other natural notions suitable
for processes.
AMS subject classification: 37A35, 37A15.
Keywords: Ergodic theory, entropy, examples.
Abbreviated title: Entropy dimensions.
∗
The authors were supported in part by a joint CNRS/KOSEF Cooperative Research Grant between
Korea and France. The second author was also supported in part by grant BK 21. The second author would
like to thank Korea Institute for Advanced Study for the pleasant stay while this work was completed.
1
Let (X, B, µ, σ, P ) be a process, where σ denotes an action of a group G, and P =
{P0 , . . . , Pk−1 } denotes a (finite, measurable) partition of X. In the study of a general group
action, subgroup actions play an important role: if a G-action has positive entropy, it is not
hard to see that every non-cocompact subgroup action has infinite entropy (see for example
[3]). In the case of a Z 2 -action generated by two commuting maps, say T and S, if either
h(T ) or h(S) is finite, the entropy of the Z 2 -action is 0. Hence it is increasingly important
to study systems of entropy zero, as they may give rise to interesting subgroup actions, and
to classify them up to measure-theoretic isomorphism. One way to achieve this goal is to
look at the amount of determinism in the system, in a more precise way that is given by
the mere knowledge of the entropy. Several refinements of the notion of entropy have been
introduced by various authors, such as the slow entropy [5], the measure-theoretic complexity
[4], the entropy convergence rates [1]; following a suggestion of J. Milnor, we propose a new
notion, the entropy dimension; though it seems most promising for actions of groups like Z p ,
for simplicity we develop here the basic definitions and examples in the case of Z -actions.
1
Growth rates and names
A first tentative way to define an entropy dimension would be to define
H
D (P ) = sup{0 ≤ α ≤ 1; lim sup
1
H(∨n0 T −g P ) > 0}.
nα
This can be generalized to Z k by taking a joint on a suitable part Gn instead of the interval
[0, n], and letting α vary from 0 to k. However, this does not define an isomorphism invariant,
H
as the following proposition implies that supP D (P ) = 1 :
H
Proposition 1 For any given P with D (P ) < 1 and any δ > 0, there exists P̃ such that
H
H
|P − P̃ | < δ, and D (P̃ ) > D (P ).
Proof
H
Let α0 = D (P ). We choose α0 < α < 1. We build a Rokhlin stack of height n1 such that
nα−1
≤ 2−L for a very large L. We may ensure that the distribution of the columns on the
1
1 −1
base level B0 of the stack is the same as the distribution of ∨ni=0
T i P . We divide each column
α
into 2n1 subcolumns of equal measure and change the partition P into P̃ 1 on the first nα1
levels from the bottom, so that each subcolumn has a different P̃ 1 -[0, nα1 )-name. For x and
y in B0 , their P̃ 1 -[0, n1 )-names may agree if their P̃ 1 -[nα1 , n1 )-names are the same, so the
α
number of different P̃ 1 -[0, n1 )-names may be smaller than 2n1 times the number of columns.
α
α
However, there are at least 2n1 different P̃ 1 -[0, n1 )-names, each one of measure at most 2−n1 .
Hence
α
1 −1
2nX
1
1
1
1
n1 −1 i 1
H(∨
T
P̃
)
≥
log nα ± i=0
α
α
nα
n1
n1 i=0 2 1
2 1
= log 2 ± 1
where comes from the error set. Also
nα
1 −1
1
1 X 2X
λ
λ
n1 −1 i 1
H(∨
T
P̃
)
≤
log nα ± i=0
α
α
nα
n1
n1 λ i=0 2 1
2 1
=
X
X
1
(−
λ
log
λ
+
nα1 log 2) ± ) = log 2 ± α
n1
λ
λ
where λ denotes the measure of a column and we sum over all columns.
We note that |P − P̃ 1 | < 2−L . Let E1 denote the nα1 levels where P and P̃ 1 may differ.
We repeat this for Rokhlin stacks of height nk where nkα−1 ≤ 2−L−k for k = 2, 3, . . .. In the
k-th Rokhlin stack, we choose nαk many levels in each column and change P̃ k−1 to P̃ k on
α
these levels so that there are at least 2nk many different names for each column. we choose
k
k−1
these levels so that their union Ek is disjoint from ∪k−1
| < 2−L−k , and
i=1 Ei . Thus |P̃ − P̃
H
we can define P̃ = lim P̃ k . And we have D (P̃ ) ≥ α. Note also that |P − P̃ | < 2−L+1 , thus
P̃ can be chosen arbitrarily close to P . And since each P̃k is measurable with respect to
the σ-algebra generated by P , so is P̃ ; if P̃ generates a factor σ-algebra, we can modify it
further so that it generates the whole σ-algebra. ♣
Remark
It is possible to define D using lower instead of upper limits. Note that if α = 1 the
construction of P̃ is not possible.
For a point x in X, the P -name of x is the sequence P (x) where Pi (x) = l whenever
σ i (x) is in Pl ; we denote by P[0,n) (x) the sequence P0 (x) . . . Pn−1 (x). Between P[0,n) (x) and
P[0,n) (y), there is the natural Hamming distance, counting the ration of different coordinates
in the names: for two sequences a = (a1 , ...ak ) and b = (b1 , ...bk ) over a finite alphabet, we
recall that
1
d(a, b) = #{i; ai 6= bi }.
k
We can define a complexity dimension for a process by
1
log #{different P − [0, n) − names} > 0},
nα
1
D0 (P ) = sup{0 ≤ α ≤ 1; lim inf α log #{different P − [0, n) − names} > 0}.
n
However it is easy to see, as in the previous case, that this is not an isomorphism invariant.
Hence, instead of counting names, we should use the number of d-balls around names.
D0 (P ) = sup{0 ≤ α ≤ 1; lim sup
2
Entropy dimensions and subgroup actions
Definition 2 For a point x ∈ X, we define
B(x, n, ) = {y ∈ X; d(P[0,n) (x), P[0,n) (y)) < }.
And let K(n, ) be the smallest number K such that there exists a subset of X of measure at
least 1 − covered by at most K balls B(x, n, ). Then
D(P, ) = sup{0 ≤ α ≤ 1; lim sup
2
1
log K(n, ) > 0},
nα
D(P ) = lim D(P, ),
→0
D = sup d(P ).
P
Similarly
1
log K(n, ) > 0},
nα
D(P ) = lim D(P, ),
D(P, ) = sup{0 ≤ α ≤ 1; lim inf
→0
D = sup d(P ).
P
We call D, resp. D, the upper, resp. lower, entropy dimension of the system (X, B, µ, σ). If
D = D, we just call it the entropy dimension and denote it by D.
Note that for a Z -action, the entropy dimension may be 1 while the entropy is 0.
It is a straightforward consequence of our definition, proved by the same proof as Corollary 1 in [4], that
D(P ) = D
when P is a generating partition.
We want to investigate the relation between the entropy dimension and the entropy of
subgroup actions, particularly in the case of Z 2 : if one of the directions has positive entropy,
then K(n, ) grows at least at the rate of ecn and the lower entropy dimension is at least
one. Hence, if D < 1, then h(v) = 0 for every direction v, and, moreover, the cone entropy
[2] has the property that hc (v) = h(v) = 0. The converse is not true: Katok and Thouvenot
[5] provide an example where the upper entropy dimension is arbitrarily close to 2 while the
directional entropy is 0 for almost all directions; note that in this example the upper and
lower entropy dimensisons do not agree.
We recall that there exists an example in [7] where h(σ (1,0) ) > 0 while all the remaining
directional entropies (including the irrational directions) are 0; this Z 2 -action has clearly
entropy dimension equal to 1. In the well-known ewample of Ledrappier ([6]), the entropy
dimension is 1 and every directional entropy is positive. By making a direct product of
countably many copies of that example, we can build a Z 2 -action whose entropy dimension
is 1 and every direction has infinite entropy, because of the following lemma, which holds
also for countable products:
Lemma 3
D(σ × τ ) = max(D(σ), D(τ )).
Proof
The entropy dimension of σ × τ may be computed by taking only partitions of the form
P × Q. But then for these partitions B((x, y), n, 2) contains B(x, n, ) × B(y, n, ) (respectively for P and Q) and is included in B(x, n, 2) × B(y, n, 2), which yields the result. ♣
3
3
Examples of entropy dimensions
We define inductively a family of blocks Bn,i , 1 ≤ i ≤ bn , in the following way; given two
sequences of positive integers en and rn :
• b0 = k, B0,i = i − 1, 1 ≤ i ≤ k,
• bn+1 = (bn )en ,
0
• the Bn,i
, 1 ≤ i ≤ bn+1 are all the possible concatenations of en blocks Bn,i ,
• for each 1 ≤ i ≤ bn+1 , Bn+1,i is a concatenation of rn blocks Bn,i .
0
Let hn be the length of the Bn,i , h0n be the length of the Bn,i
.
We can thus define a topological system as the shift on the set of sequences {xn , n ∈ ZZ}
such that for evry s < t there exists n and i such that xs . . . xt is a subword of Bn,i . We put
an invariant measure on it by giving to each block Bn,i the measure b1n . We denote by P the
natural partition in k sets given by the zero coordinate.
The above construction is well known to ergodic theory specialists, and a generalization of
it to Z 2 -actions is used in [5]; however, even its one-dimensional version can yield new types
of counter-examples. This system will be referred in the sequel as the standard example.
Proposition 4 There is a choice of parameters such that the standard example satisfies
D = 1,
D = 0.
Proof
Lower limit.
For any , K(hn+1 , ) is smaller than the total number of P − [0, hn+1 )-names. The possible names of length hn+1 are all the Wn+1 (a, i, j) where, for 0 ≤ a ≤ hn+1 , 1 ≤ i ≤ bn+1 ,
1 ≤ j ≤ bn+1 , Wn+1 (a, i, j) is the suffix of length a of Bn+1,i followed by the prefix of
length hn+1 − a of Bn+1,j . Hence their numbers is at most hn+1 b2n+1 , with bn+1 as above and
hn+1 = e0 . . . en−1 r0 . . . rn . If, e0 , . . . ,en , r0 , . . . , rn−1 being fixed, we choose rn large enough,
we shall have log K(hn+1 , ) ≤ (hn+1 )δn for any given sequence δn .
Upper limit
¯
Let L0n () be the number of -d-balls
than can be made with blocks Bn0 . Note that, on
an alphabet of k letters, for a! given word w of length m, the number of words w0 with
m
d(w, w0 ) < is at most
k m ≤ k mg() for some g() → 0 when → 0. In the above
m
construction, the number of different blocks Bn0 is bn+1 = k e0 ...en . As in every of these blocks
0
the repetitions occur exactly at the same places, for
a given word Bn,i
, the number of words
!
e0 . . . en
0
0
0
Bn,j
with d(Bn,i
, Bn,j
) < is at most
k e0 ...en . Hence
e0 . . . en L0n () ≥ k e0 ...en −g() .
4
As all different blocks are given the same measure, we have
K(h0n , ) ≥ (1 − )Ln ().
As h0n = e0 . . . en−1 r0 . . . rn−1 , if, e0 , . . . ,en−1 , r0 , . . . , rn−1 being fixed, we choose en large
enough, we shall have log K(h0n , ) ≥ (h0n )1−δn for any given sequence δn .♣
Proposition 5 For any 0 < α < 1, there is a choice of parameters such that the standard
example satisfies D = α.
Proof
We make the proof for α = 21 . We define a sequence ln by choosing a very large l1 , then
1
1
2(n−1)
2(n−1)
ln = [ln−1
][ln−1
]ln−1 ; then, in the standard construction, starting from the two 0-blocks
1
√
2(n−1)
0 and 1, we put e0 = r0 = [ l1 ], and, for n ≥ 1, en = rn = [ln−1
]. [x] denotes the integer
part of x, but in the following computations we shall assimilate a large enough x with its
integer part.
Then, the lower limit is reached along the sequence {hn+1 = ln } and the upper limit
1+
1
along the sequence h0n = {ln−12(n−1) }.
Lower limit
As in the second part of the proof of the last proposition, K(hn+1 , is at least (1−) times the
¯
number Ln+1 () of -d-balls
than can be made with blocks Bn+1 . Because of the repetitions,
and the computation in the last proposition
Ln+1 () ≥ L0n () ≥ 2e0 ...en −g() .
So we have only to compute
1
lim √ log(2e0 ...en ) =
n→+∞
ln
q q
q
1
1
1
√ √
n−1
1
l1 l1 l22 l33 ... ln−1
lim √ log 2
=
n→+∞
ln
1 q
ln log 2 = log 2.
lim √
n→+∞
ln
Hence D ≥ 12 .
Upper limit
As in the first part of the proof of the last proposition, K(h0n , ) is smaller than the total
number of P − [0, h0n )-names, and this is at most b2n+1 h0n . We take some b > 21 ;
lim sup
n→+∞
1
log b2n+1 h0n =
0
b
(hn )
2 lim sup
n→+∞
1
log bn+1 =
(h0n )b
5
2 lim sup
n→+∞
√ √
l 1 l1
1
1+ 1
(ln−12(n−1) )b
log 2
r
2 lim sup
n→+∞
q q
1
l33 ...
q
1
n−1
ln−1
=
1
1+
ln−12(n−1)
1+
1
l22
1
= 0.
(ln−12(n−1) )b
Hence D ≤ 12 , which gives what we claimed.
α
1−α
n−1
n−1
The general case (for a given α) follows with the same proof, by taking ln = ln−1 [ln−1
][ln−1
],
α
1−α
n−1
n−1
en = [ln−1
], rn = [ln−1
].♣
The above examples can be generalized to Z 2 -actions; by alternating repetitions and independent stacking, we can build an example whose entropy dimension is any given 0 ≤ α ≤ 2.
In [4], where the rate of growth of K(n, ) is used to define the so-called measure-theoretic
complexity, it is asked whether this growth rate can be unbounded but smalller than O(n)
(its topological version for symbolic systems, the symbolic complexity has to be bounded if
it is smaller than n). Our class of examples allows to answer this question; note that the
proofs are slightly more involved as we are dealing with sub-exponnetial growths:
Proposition 6 For any given function φ growing to infinity with n, there is a choice of
parameters such that the standard example satisfies, for every fixed small enough,
K(n, ) → +∞
with n, but
K(n, ) ≤ φ(n)
for all n.
Proof
Upper bounds
¯
We give upper bounds for K 0 ≥ k, where K 0 (n, ) is the smallest number of -d-balls
of
names of length n necessary to cover a proportion of the space of measure 1. We look at K 0
at the end of its times of maximal growth, namely K 0 (h0n , ). The possible words of length
h0n are all the Wn0 (a, i, j) where, for 0 ≤ a ≤ h0n − 1, 1 ≤ i ≤ bn+1 , 1 ≤ j ≤ bn+1 , Wn0 (a, i, j)
0
0
is the suffix of length a of Bn,i
followed by the prefix of length h0n − a of Bn,j
. Each one of
0
¯
these words is at a (d) distance at most of some Wn (as , is , js ) for 1 ≤ s ≤ K 0 (h0n , ).
We look now at words of length hn+1 ; they are all the Wn (a, i, j) where, for 0 ≤ a ≤
hn+1 − 1, 1 ≤ i ≤ bn+1 , 1 ≤ j ≤ bn+1 , Wn (a, i, j) is the suffix of length a of Bn+1,i followed
by the prefix of length hn + 1 − a of Bn+1,j . Hence for 0 ≤ t ≤ rn − 1, 0 ≤ a ≤ hn+1 − 1,
1 ≤ i ≤ bn+1 , 1 ≤ j ≤ bn+1 ,
Wn (a + th0n , i, j) = W (a, i, i)t W (a, i, j)W (a, j, j)rn −t−1 .
6
Each one of these will be at a distance at most + r1n of some W (as , is , is )t+1 W (as , js , js )rn −t−1 ;
0
0
and, for fixed s, W (as , is , is )t+1 W (as , js , js )rn −t−1 and W (as , is , is )t +1 W (as , js , js )rn −t −1 are
0
|
at a distance at most |t−t
. Hence, for a given sequence vn , we have
rn
K 0 (hn+1 , (1 +
K 0 (h0n , )2
1
+ vn )) ≤
.
rn
vn
Then, during the stage of independent stacking, a straightforward computation gives that
K 0 (h0n+1 , ) ≤ K 0 (hn+1 , )en .
P
1 < +∞; we choose any sequence vn such that
If we fix the sequence en , and suppose
rn
P
vn < +∞; then, if we choose rn large enough in terms of K 0 (h0n , ), h0n , and en , we get
that K 0 (h0n+1 , 2) is smaller than φ(h0n+1 ), and this is true a fortiori for other values.
Lower bounds
¯
We shall show that K(n, ) → +∞ with n. For this, let Ln () be the number of -d-balls
¯
than can be made with blocks Bn , and L0n () be the number of -d-balls
that can be made
0
with blocks Bn . During the repetition stage, we have
Ln () ≥ L0n−1 ().
¯
Then, during the independent stage, we start from L = Ln () blocks which are -d-separated;
en
we call them Bn,s1 , ... , Bn,sL . Then, if en is a multiple of L, the 2L blocks Bn,si , 1 ≤ i ≤ L,
en
en
en
en
en
1
L
L
L
L
L
¯
and Bn,s
i Bn,si+1 ...Bn,sL Bn,s1 ...Bn,si−1 , 1 ≤ i ≤ L, are (1 − L )-d-separated.
Thus whenever en is large compared to Ln () we have
Ln+1 ((1 −
1
)) ≥ 2Ln ()
Ln ()
and hence Ln ( 2 ) tends to infinity with n; and, because of the structure of the names and
the fact that each block Bn,i has the same measure for fixed n, we get that K(hn , ) tends
to infinity with n. ♣
Remarks
To make our examples weakly mixing, it is enough to place a spacer between two consecutive
blocks at each repetition stage.
It is easy to see that all our examples satisfy a form of Shannon-McMillan-Breiman
theorem (indeed, all atoms have the same measure); in a forthcoming paper, we shall give
examples which do not satisfy it.
References
[1] F. BLUME: Possible rates of entropy convergence, Ergodic Theory Dynam. Systems 17
(1997), no. 1, 45–70.
7
[2] R. BURTON, K. K. PARK: Spatial determinism for a Z 2 -action, preprint.
[3] J.-P. CONZE: Entropie d’un groupe abélien de transformations, (in French), Z.
Wahrscheinlichkeitstheorie und Verw. Gebiete 25 (1972/73), 11–30.
[4] S. FERENCZI: Measure-theoretic complexity of ergodic systems, Israe̋l J. Math. 100
(1997), 189-207.
[5] A. KATOK, J.-P. THOUVENOT: Slow entropy type invariants and smooth realization of commuting measure-preserving transformations, Ann. Inst. H. Poincaré Probab.
Statist. 33 (1997), no. 3, 323–338.
[6] F. LEDRAPPIER: Un champ markovien peut être d’entropie nulle et mélangeant, (in
French), C. R. Acad. Sci. Paris Sr. A-B 287 (1978), no. 7, A561–A563.
[7] K. K. PARK: On directional entropy functions, Israe̋l J. Math. 113 (1999), 243–267.
8