Progress in Probability and Statistics Vol. 8
Edited by Peter Huber Murray Rosenblatt
Birkhauser
Boston Basel Stuttga...
20 downloads
637 Views
3MB Size
Report
This content was uploaded by our users and we assume good faith they have the permission to share this book. If you own the copyright to this book and it is wrongfully on our website, we offer a simple DMCA procedure to remove your content from our site. Start by pressing the button below!
Report copyright / DMCA form
Progress in Probability and Statistics Vol. 8
Edited by Peter Huber Murray Rosenblatt
Birkhauser
Boston Basel Stuttgart
Philippe Bougerol Jean Lacroix
Products of Random Matrices with Applications to Operators
Editors Philippe Bougerol Universite Paris 7 UER de Mathematiques 2 Place Jussieu F-75005 Paris/France
Jean Lacroix Universite de Paris XIII Departement de Mathematiques F-93430 Villetaneuse/France
Library of Congress Cataloging in Publication Data Bougerol, Philippe, 1951Products of random matrices with applications to Schrodinger operators.
(Progress in probability and statistics ; v. 8) Bibliography: p. 1. Stochastic matrices. 2. Schrodinger operator. II. Title. I. Lacroix, Jean, 1944- . III. Series.
QA184.B68 1985 ISBN 0-8176-3324-3
512'.5
85-15695
CIP-Kurztitelaufnahme der Deutschen Bibliothek Bougerol, Philippe: Products of random matrices with applications to Schrodinger operators / Philippe Bougerol ; Jean Lacroix. - Boston ; Basel ; Stuttgart : Birkhauser, 1985.
(Progress in probability and statistics ; Vol. 8) ISBN 3-7643-3324-3 (Stuttgart ... ISBN 0-8176-3324-3 (Boston)
NE: Lacroix, Jean:; GT
© 1985 Birkhauser Boston, Inc. Printed in Germany ISBN 0-8176-3324-3 ISBN 3-7643-3324-3
CONTENTS
"LIMIT THEOREMS FOR PRODUCTS OF RANDOM MATRICES"
PART A
INTRODUCTION
CHAPTER I
1
THE UPPER LYAPUNOV EXPONENT
-
.
5 5
1. Notation
2. The upper Lyapunov exponent
6
3. Cocycles
8
4. The theorem of Furstenberg and Kesten
11
5. Exercises
13
CHAPTER II
-
17
MATRICES OF ORDER TWO
1. The set-up
17
.
2. Two basic lemmas
19
3. Contraction properties
24
4. Furstenberg's theorem
30
5. Some simple examples
33
6. Exercises
.
36
7. Complements
.
38
CHAPTER III - CONTRACTION PROPERTIES
.
43
1. Contracting sets
44
2. Strong irreducibility
48
3. A key property
4. Contracting action on direction
50
P(]Rd) and convergence in 55
v
5. Lyapunov exponents
60
6. Comparison of the top Lyapunov exponents and Furstenberg's theorem
64
7. Complements. The irreducible case
68
.
CHAPTER IV
-
.
COMPARISON OF LYAPUNOV EXPONENTS AND BOUNDARIES
77
.
1. A criterion ensuring that Lyapunov exponents
are distinct
77
2. Some examples
81
3. The case of symplectic matrices
87
4. p-boundaries
93
CHAPTER V
CENTRAL LIMIT THEOREMS AND RELATED RESULTS
-
1. Introduction
.
101 101
2. Exponential convergence to the invariant
measure
103
3. A lemma of perturbation theory .
111
4. The Fourier-Laplace transform near 0
116
5. Central limit theorem
121
.
6. Large deviations
129
y
7. Convergence to
134
.
p
8. Convergence in distribution without normalization
135
9. Complements equations
140
CHAPTER VI
-
:
linear stochastic differential
PROPERTIES OF THE INVARIANT MEASURE AND APPLICATIONS
145
1. Convergence in the Iwasawa decomposition
147
2. Limit theorems for the coefficients
155
3. Behaviour of the rows
159
4. Regularity of the invariant measure
161
5. An example : random continued fractions
166
SUGGESTIONS FOR FURTHER READINGS
173
BIBLIOGRAPHY
175
vi
"RANDOM SCHRODINGER OPERATORS"
PART B
INTRODUCTION
CHAPTER I
183
THE DETERMINISTIC SCHRODINGER OPERATOR
-
.
187
1. The difference equation. Hyperbolic structures
187
2. Self adjointness of H. Spectral properties
190
.
3. Slowly increasing generalized eigenfunctions
195
4. Approximations of the spectral measure .
196
5. The pure point spectrum. A criterion
200
6. Singularity of the spectrum
202
CHAPTER II
ERGODIC SCHRODINGER OPERATORS
-
.
"
.
205
1. Definition and examples
205
2. General spectral properties
206
3. The Lyapunov exponent in the general ergodic case
209
4. The Lyapunov exponent in the independent case
211
5. Absence of absolutely continuous spectrum
221
6. Distribution of states. Thouless formula
224
7. The pure point spectrum. Kotani's criterion
232
8. Asymptotic properties of the conductance in the disordered wire
234
THE PURE POINT SPECTRUM
237
1. The pure point spectrum. First proof
238
2. The Laplace transform on
240
CHAPTER III
-
Sl(2,]R)
3. The pure point spectrum. Second proof
247
4. The density of states
250
CHAPTER IV
-
.
SCHRODINGER OPERATORS IN A STRIP
.
253
1. The deterministic Schrodinger operator in
a strip
253
2. Ergodic Schrodinger operators in a strip
259
3. Lyapunov exponents in the independent case. The pure point spectrum (first proof)
2 62
4. The Laplace transform on
267
Sp(i.,IR)
.
5. The pure point spectrum, second proof
vii
272
APPENDIX
275
BIBLIOGRAPHY
277
viii
PREFACE
This book presents two closely related series of lectures.
Part A, due to P. Bougerol, is an introduction to the works of Furstenberg, Guivarc'h, Le Page and Raugi on products of random matrices. Only invertible independent identically distributed random matrices satisfying an irreducibility condition are considered. The purpose is to prove in detail the analogues of the classical limit theorem (e.g. law of large numbers, central limit theorem). This part is based on a course given at the University of Paris 7 in 1983.
Part B, due to J. Lacroix, deals with the spectral theory of random Schrodinger operators, where the products of random matrices play a crucial role. It presents a rigorous and unified treatment of the main known results in the one-dimensional discrete case. Since we are aware that some readers are mainly interested in Schrodinger operators, any notion or result needed from part A is clearly restated (but of course not proved again !).
The book is self-contained and should be accessible to readers with minimal background. Since we feel that these topics deserve a large audience we have "tried" to write it English. It is not sure that we have succeeded and we beg the indulgence of the native speaker.
It is our pleasure to thank Yves Guivarc'h, Francois Ledrappier Emile Le Page, AlbertRaugi, Jose de Sam Lazaro and Bernard Souillard for enlightening conversations concerning the material presented here and for their encouragements.
ix
Finally we thank Joelle Navarron and Catherine Simon for their typing.
Paris, May 1985.
x
PART
A
LIMIT THEOREMS FOR PRODUCTS OF RANDOM MATRICES
INTRODUCTION
This part is devoted to limit theorems for products of i.i.d. invertible random matrices. The subject matter, initiated by Bellman, was fully developed by Furstenberg, Guivarc'h, Kesten, Le Page and Raugi. This text is intented to serve as an introduction to their work.
We have chosen to keep the level as elementary as possible. This had sometimes led us to write lengthy proofs when shorter ones are available and to omit some important topics. On the other hand the text is self-contained and should be accessible to readers familiar with probability theory as usually developed at graduate level. In particular, no prior knowledge of group theory is assumed.
Let us roughly describe our general line of approach. We consider a sequence matrices of order We set
So = Id
d
Y1,Y2,...
of invertible random
which are independent and identically distributed.
S1 = Y1,....Sn = Yn...Y2Y1.
,
Our purpose is to prove that under an irreducibility condition, the sequence Log 1Sn xll
satisfies, for any
x # 0
;
n = 1,2,...
in ]Rd,analogues of the classical limit
theorems for sums of i.i.d. random variables (e.g. law of large numbers, central limit theorem, ...).
The strategy we shall adopt is based on the following observation.
For any class of F
:
x
x # 0
in
]Rd, let
in the projective space
Gl(d,]R) x P(]Rd) -> ]R
x
defined by 1
denote its direction (i.e. the
P(]Rd)). Consider the function
2
_
IlYxil
F(Y,x) = Log
.
IIx 11
Then IIS n
Log
n-1
xil =
II S
1+1
Log
E
IISixII
i=o n-1
IIXII =
E
i=o
x11
S. x Log IIYi+1
1
II
IISixII
n-1 E
i=o Therefore
Log IISnxII
of the Markov chain
(Y
i+11
F(Yi+1' Six)
can be written as a simple functional
Six)
on
Gl(d,IR) X P(1Rd)
.
Taking into
account our irreducibility assumption, the first conclusion we can draw is the law of large numbers. Namely, there exists a real the upper Lyapunov exponent, such that for any
n-
, called
a. s.
,
lim n Log II Sn xII = -Y1
yl
x # 0
The other results require a detailed study of the Markov chain (Y1+11`6.1 x)
.
Its essential properties will be derived from the fact
that, loosely speaking, the random matrices
Sn
have asymptocally a
contracting action on the directions. To be more precise, let the natural angular distance on
6
be
P(JRd)
6(x,y) = Isin{angle(x,y)}I
We shall prove, following Guivarc'h and Raugi [34], that under fairly weak assumptions,
lim sup n Log 6(Snx, Sny) < 0
a. s.
(1)
n-
f or any fixed x and y in
P (]Rd)
This property is of primary importance. For instance it yields easily that the Markov chain
(Yi+1' Six)
has a unique invariant
probability measure. More crucially we shall see, following Le Page [49]
,
that (1) is the cornerstone on which rest the proofs of the
central limit theorem, the estimate for large deviations and many further results. We might say that, in some sense, this inequality replaces Doeblin's condition under which the precise limit theorems for functionalsof Markov chains are usually available.
Let
yl > y2 > with the sequence (Yn)
.
> Yd Since
be the Lyapunov exponents associated
n
lim sup n Log d (Sn X , our first task is to show that
Sn Y)
< Y2 - Y1
Y, > Y2 .
The main purpose of the first three chapters is to derive the inequality
from a careful analysis of the qualitative
yl > Y2
behaviour of the Markov chain
(Snx).
In Chapter I we introduce the basic properties of the upper Lyapunov exponent
yl
.
In Chapter II we restrict ourselves to
2 X 2
matrices. For
pedagogical purpose we first develop in this simple setting the general argument leading to a proof of this inequality. Most of the results which are needed in part B can already be found here.
Chapter III treats the case of matrices of arbitrary order.
In Chapter IV we digress from the main line. We apply the preceding results to the study of all the Lyapunov exponents. We also briefly present the link between these exponents and the boundary theory of Furstenberg.
In Chapter V we derive the main limit theorem on from the inequality
y1 > Y2
.
Log IISnxll
Further results, such as limit theorems
for the coefficients of the matrices
S
n
,
are proved in Chapter VI.
All the main results proved in this text come from Furstenberg
[21]
,
Guivarc'h and Raugi [341
and Le Page [49] . We hope that our
exposition entices the reader to go back to these profound original works.
Despite its importance, we have chosen not to consider Osseledec's theorem. The reason is twofold. Firstly we have tried to keep the prerequisites to a minimum and to give a self-contained account of the subject. Secondly Ledrappier has already given in r46]
a beautiful treatment of the applications of this theorem to products of i.i.d. random matrices. We felt there was nothing to gain by a
4
reproduction of this material. In our opinion, the reader who wants to have a full picture of random products has to read Ledrappier's monograph. It will then be an easy and useful exercise to him to check how some of our proofs can be shortened by making use of his results.
We have not considered positive matrices (except in some exercises). We understand that Joel Cohen is writing a book on this subject.
For the sake of simplicity of notation we have restricted ourselves to matrices with real entries. But all the statements are also true in the complex case (the proofs carry over immediately to this case by replacing everywhere
IR by
M)
.
Some chapters contain "complements" sections. They develop some additional material which is not used elsewhere and the reader can skip them. Some of their proofs are only outlined and may be of a more advanced level.
CHAPTER I
THE UPPER LYAPUNOV EXPONENT
In this chapter we define the upper Lyapunov exponent
y
which
gives the exponential rate of growth of the norm of products of
independent identically distributed (i.i.d.) random matrices. In order to prove the analogue of the law of large numbers we develop some basic results on G-spaces which will often be used in the sequel.
I.1. Notation
We shall write
M(d,]R) for the set of
d x d
is the set of invertible elements of
entries. Gl(d,IR)
M
the set of matrices with determinant one. If in
matrices with real
is in
M(d,]R), Sl(d,]R)
M(d,]R) and
x
]Rd, Mx will be the vector equal to the product of the matrix M
by the column vector
x
.
The transpose of
M
is
M
In general the results we present do not depend on a particular norm chosen on
]Rd or
M(d,IR).
For convenience we shall use, unless
tho contrary is explicitely stated, for
x e ]Rd and
M c M(d,]R)
d jxII
E
_
i=1 IM11
= Sup{ JIM-11
:
x E IRd
jjxjj = 1}
Similar definitions hold for matrices with complex entries.
5
6
It will be understood that the random variables are defined on a
probability space random variable
(0, A, 1P) , 1E (X) will be the expectation of a = J X d1P, and if
X, ]E(X)
F
is a sub
a-algebra of
A, 1EF(X) will be the conditional expectation with respect to F A topological semigroup is a topological set associative product is defined, the map
T
on which an
(M1,M2) -> M M I
T
being continuous. For instance A topological group is a group
mappings
(g,h) -> gh
G
T2
to
with a topology for which the
g -> g-1
and
from 2
is a topological semigroup.
M(d,]R)
are continuous.
I.2. The upper Lyapunov exponent
{Yn, n >_ 0 of i.i.d. random matrices with
Consider a sequence common distribution
p
.
IISnII < Therefore if Log+ IISnII
]E(Log
Sn = Yn...Y1. For the chosen norm,
Let
is integrable. If
n,p ?
]E (Log II S n 11 )
for
Sup(f,0))
IE(Log IIYn+p...Yn+1II) +]E(Log IIYn.Y1II)
Thus the sequence fIE(Log IISn II),
n
f
then
1
IE (Log
1
IIYI II
is finite (we write
IIYIII)
IE(Log IISn+p II)
...
IYn_, II
IIYn II
IIS
n
1
1}
1)
IE (Log IISnII) is subadditive and
converges to inf 1m]E (Log I I S m II ) in
IR U
m?1
(see exercise 5.2).
DEFINITION 2.1.
If 1E(Log+ IIY1II) < 00 , the upper Lyapunov exponent
associated with
pi
is the element
Y=1 n
n
y
of IR U {_co}
defined by
]E(Log IIYn...Y1 I1)
Since all norms on the finite dimensional vector space are equivalent, (when
Y1
y
M(d,]R)
is independent of the chosen norm. We shall prove
is invertible) the stronger result that almost surely (a.s.)
Y =1imnLogllYn...Y1I n-_
7
Although this turns out to be an easy consequence of Kingman's subadditive ergodic theorem (see e.g. Ledrappier [46]) we shall follow the original proof of Furstenberg and Kesten [25]. The reason is twofold. On one hand it is a good illustration of the techniques we shall use. On the other hand it gives us some information that will be needed later. Let us first describe two examples.
Suppose that each matrix Y
Example 2.2.
is a diagonal matrix
n
diag(a1(n),...,ad(n)). This means that the entries
(Yn )1,J It is clear that
_ I
0
if
i # j
ai (n)
if
i=j
y = Sup1E(Loglai(1)l)
satisfy
(Yn)i,J
.
and
i
y = lim
Log IIYn . .Yl II n
by the usual law of large numbers.
Example 2.3.
The preceding example is degenerated. Far more interesting
is the following due to Cohen and Newman [15]. Let G1(d,1R)
Y1,Y2,...
be a sequence of i.i.d. random elements in
such that the distribution of
IIY X ll 1
x # 0
in 1Rd. For each vector
x
does not depend upon
Ilxll
0, the real random variables Y
Log IIY X 11,...,U
U,
1
11- 11
= Log IIY
n
are independent and identically distributed a-algebra generated by
n1 :
.Y x
n-1 n-
for, if
Il,
1
...
lYn-1...Y1xil
F
n
is the
U1,...,Un, for any Borel bounded function
f
on 1R F 1E
{f(Un+1)} = 1E{f(Log llYn+lyll)} = 1E{f(U1)}
Yn...Y1x where
y =
.
Remark that
IIYn. . .Y1x ll
Log IIYn...Yjxll = Un + Un-1 + ... + U1 + Log IIxII Therefore we may apply the classical limit theorems for sums of i.i.d. random variables in order to obtain results on the sequence
3
{Log IY....Y1xII} . If 1E(Log+ IIY1II) law of large numbers
IE(U+) < - and, by the usual
,
0) , G
M.
x
the set of
,
(i.e. matrices
M(d,1R)
with nonnegative coordinates. With the
S
and
S+
defines an additive cocycle on
a
GxS Let
DEFINITION 3.3.
p * v
be a topoZogicaZ semigroup acting on
G
is a probabitity measure on
(resp. v)
the distribution on
which satisfies
B
f(x)d(p * v)(x) =
G
for all bounded BoreZ function
p * A
on
G. Note that
distribution on
and
a
distribution
If v
Let
p
G
,
the ordinary
f(gh)du(g)da(h)
(p * A) * v = p * (A * v)
G
if
v
is a
be a topological semigroup acting on the space G x B
.
Consider a sequence
of independent random elements of
G
with common
B
such that
.
p-invariant distribution on
is a
o(g,x)Idp(g)dv(x)
JJ
JI
an additive cocycle on
{Yn, n ? 1}
.
B.
PROPOSITION 3.4. B
.
is defined by
Jfd(p * A) = f
B
are two probability measures on
A
convolution product
for
on
f
(1)
B
p-invariant if p * v = v
is
v
and
p
If
J
I
B
We say that
B. If p
we denote by
(resp. on B)
G
is finite, then
if
lim n a(Yn(w) ... Y1(M ), x) n-
exists for 1P®v-almost all Proof
(w,x)
and in L1(W®v)
.
Without loss of generality we may suppose that Yn e G} , A
0 = {w = (Yn)n:21
a-algebra generated by the
being the
'
coordinate maps Y
's i
Yn
and
IP
the probability measure for which the
are independent with distribution
p
.
1G
Let E = Q x B
and
Q=IPOV
.
Define
0
: E -> E
by
0(((Yn)n>l,x)) = If
is a Borel subset in
A0
Q{((o,x) E Q x B
and
B
Borel subsets in
G,
(A1 X A2 x ... x An x ...) x Ao}
0((w,x)) e
;
Al,A2,...,An
= (1P ®V) {y
c A1,...,Yn+1 c
= (IP0V) {Y1
F- Al,...,Yn c An...,x e Ao}
E Ao}
= Q{(A1 X A2 x ... x An x ...) x Ao) This proves that
preserves
0
For each integer
p
Q
,
0P(((Yn)n>1.x)) = ((Yn+p)n>1, so that if we set
F(w,x) = G(Y1,x)
for
w = (Y.).,,,1, x e B, then D(Yn-l*** YI'X)
G(Yn...Yi,x) = O(Yn'(Yn-l'** n =
E
.. Y
a(Y , (Y P_1
p
p=1
1
n =
E
F(0p((w,x)))
P=1
Therefore Birkhoff's ergodic theorem (see e.g. Breiman Theorem 6.2.1) implies that
Q(Yn...Y1,x) = n n converges
distributions on power of LEMMA 3.5.
F(0p(w,x)))
L1
is separable and compact there exist
B
If
a.s. and in
1P R V
E
P=1
B
:
(p1
p
= u,
Let
indeed if pn denotes the u2 = p * p, ...) , we have
be a compact separable
B
11-invariant
nth
G-space and
convolution
p
be a
distribution on G For any distribution m on B , each limit point n p1 * of {! E m , n ? 1) is a p-invariant distribution on B .
n i=1
Proof
:
n Let
V
n
=
E
-n i=1
}11 * m
.
Since
B
is separable and compact
there exists a weakly convergent subsequence of
(vn )
and its limit,
11
say
v
,
is a probability measure. For any 1
*V n
}1
n
i+1 *
n
m
n i=1 n
lal*m+n{un+1*m-u*m}
nE i=1 Vn
+
1
{un+1
n so that, letting
n -> m
along a subsequence,
p * V = V
1.4. The Theorem of Furstenberg and Kesten
We can now prove the following result of Let
THEOREM 4.1.
1E(Log+IlY1II) < -
[25I
be i.i.d. matrices in
Y1,Y2,...
G1(d,1R). If
then with probability one
lim
Log
II
=Y
n-
Fix an integer m
Proof
p 2;0
and for every integer
n , write n = pm+q,
0 S q < m. We have
,
II < n Log 1 n Log flYn...YJ
IIYn...Ypm+1 II +
n
n
0
or all the
the sequence
are isometries. When m
Yn's
is not continuous then
has particularly simple properties and the Lyapunov
(Yn)
exponent, in this case, can be explicitely computed.
11.2. Two basic lemmas
We use the notations of 1.3
.
The following lemma due to Guivarc'h
and Raugi [34] improves a result of Furstenberg [21].
LEMMA 2.1.
Let
be a topological semigroup acting on a 2'
G
locally compact space
B
.
Consider a sequence
independent random elements of defined on
(0, A, IP). If v
then for almost all
w
G
countable
of
{Xn, n >_ 1)
with a common distribution
p
is a p-invariant distribution on
there exists a probability measure
such that the sequence (X1(w)X2(w) ... X(w)g v ,
converges weakly to respect to
A = E" n=o
vw
as
2-n-lun
n ->
n = 1)
for almost all
B
vw on
g e G with
B
20
Moreover for each bounded BoreZ function
f
f dv}
.
J f dv Proof F :
Let
f
=
1E {J
on
B
be a continuous bounded Borel real function on
B and
G -> IR be defined by F(g) = J
dv(x)
g e G
,
.
B
Set
Mn = X1X2...Xn.
If
Fn
is the 0-algebra generated by
X11 ...,Xn
then
F 1E n {F (Mn+1) } = J f (Mng) dp (g) =
du(g) dv(x)
JI
= J
-x) dV (x)
f (M
F(M ) n (we have used the independence of the
u-invariant). This shows that bounded,
F(Mn)
X.'s
and the fact that
(F(Mn), n= 1}
converges a.s. to some
1E {rf } = IF {F (M1) } = We have for any positive integers
Ff ( I
f dv
.
k, r,
{F(Mk":.) 2) + 1E {F(Mk)2}
= 1E
(F(Mkr)2} - IE {F(Mk)2)
E
IE { IF(Mk+r)-F(Mk) I2) < S
p
1E {F(Mk)2} +
E
2r Sup IF(x) I2
This yields E
1E{( IF(Mkg) - F(Mk)I2 da(g)} _
E
k=1
x
k=1
- 21E {F(Mk+r)F(Mk)}
k=1
k=1
is
and
1E { IF(Mk+r)-F(Mk) I2} = IE
Hence, by cancellation, for any
V
is a martingale. Being
.
lE {F( k+r)2}
21
E
E
2-r-1
E
E
IF(Mkg) - F(Mk)I2 dur(g)} _
1E{I
k=1 r=o
J
g-r
r-1 1E{IF(rrk+r)
2
- F(Mk)I2}
k=1 r=o So
r=o
is finite and
IF(Mkg) - F(Mk)I2
E
Sup IF(x)I2 x
r
E
F(Mkg)
converges to
rf
k=1 almost surely.
1P ® A
Choose a dense sequence
continuous functions on
A
measurable subset
of
in the space
{fq, q -> 1}
C0(B)
of
which vanish at infinity. There exists a
B
with
S2 x B
OP ® A)(A) = 1
such that for
(w,g) rA, dv(x) -> rfq(w)
.
J If
vW
is a limit point of
{Mn(w)g v ,
for the weak topology,
n >_ 1}
g
f fq dvw,g = Ffq(w)
This shows that
V
w ,g
does not depend on
We therefore denote it by
Mn(W)g V .
Vq >_
,
.
and is the limit of
g
Vw .
1
Since
1 fq dv = lE{rfq} for any bounded Borel
f
f dV
=
1E {J
f dv
1
and
Vw must be a.s. a probability measure.
For proving that Lyapunov exponents (or difference of Lyapunov exponents) are positive we shall need the following.
LEMMA 2.2.
Let
on
G x B
.
be a
B
nrydom elements of
G
G-space ,
a sequence of independent
(Yn)n =1
with distribution u and
Suppose that
v
is a
a
an additive cocycZe
W-invariant distribution on
B
that
IJ 0+(g,x) dp(g) dv(x) < - .
(i)
(ii)
For
1P ® v -almost all
(w,x)
,
I'm o(Yn(w)...Y1(w),x)
n
Then
a
is in
L1(1P 3 v)
and
(r IJ a(g,x) du(g) dv(x) > 0
such
22
Using the arguments developed in the proof of Proposition 1.3.4 this follows at once from the following lemma. Its proof is borrowed
from Dekking [16]. LEMMA 2.3.
Let
be a probability space and
(E, F, A)
a measurable transformation which preserves so that
n i=1
E
)f dA > 0
and
f E L1(da)
Proof
Jim
n-o
f+ dA < W and
J
f
A
If
.
0
:
E -> E
: E -> IR is
almost everywhere then
01 = oo
,
f
.
By the Ergodic Theorem (see Exercise 2.6) n
limn where J = {A E F
;
f
E
n-)oo
a.s.
01 = IEJ(f)
o
i=1
n
A(A A 0-1A) = 0}
Since
.
fo01
E
i=1 E`J(f)
so
>= 0
f
is actually integrable.
If f f dA = 0 then n n'
o 01 -> 0
f
E
a.s.
(3)
.
i=1 n
For each
c > 0
IE(t) = Et-, t+E]
let
,
Sn(x) =
E
(f
o 01)(x)
and,
i=1
for
m
being the Lebesgue measure on
1R,
n
IE(Si(x))}
R6(x) = m{ U i=1
By (3), for A-almost all such that, for
k > no
Rt(x) < Rn (x) + 2nd
and any
x
ISk(x)I
,
and
n Rn(x) -> 0 a.s.
.
there exists
Hence, for all
RR (x) < 26
lira
naao
O
6 > 0
0
.
Therefore
n
and
limn 1E (Rn) = 0 n-_ Note that, since
Sn o 0 = Sn+1
- S1
n+1 Rn+1 - Rn o
0
= m{ U
=
(4)
n I(Si)} E - m{ U I(Si+1-Si)} E
i=1
i=1
n+1
n+1
m{ U i=1
I
E
(S
.)} - m{ U
1
i=2
I (S.)} E
1
It is plain that
Rn+1 - Rn o 0 ? 2E
1
{ISi-S1I>2E, i = 2,...,n+1}
23
so
lE{Rn+1}- ]R{Rn o6} B 2e A{x ; ISi(x)-S(x)I > 2E,
and, since
preserves
6
i = 2,...,n+1}
A
1E{Rn+1} - 1E{Rn} B 2E
IS.(x)I > 2e, i= 1,...,n}
A{x ;
This implies that limn IE {Rn} B 2e A{x
;
> 2E
ISi(x) I
Vi >_
,
1}
nBy (4), for any
E > 0
IS.(x)I > e
A{x ;
The processes
,
{S>_.,
i
and
11
>_
{S
A{x
i
i+P-S
distribution. Therefore for any integer
0
=
Vi >_ 1}
,
1}
>_
F' p > 0
have the same
and
ISi+p(x) - Sp(x)I > E
,
Vi ->-
ISi+p(x) - SP(x)I > e
,
Vi >=
c > 0 ,
1} = 0
'
A{x ;
and
This contradicts the fact that
= m
lim S
1
for some
p} = 0
a.s.
in
Exercise 2.4.
group and
that if m gm = m
T
be a compact separable semigroup contained in a
is a distribution on
for any
contains T
Let
a probability measure on
11
in
g
T
H
If
gm = m
show that
T
.
whose support is
T
such that
T
T . Prove
11 * m = m , then
is the smallest closed group which
for any
g
in
H. Deduce from this that
is actually a group.
Exercise 2.5. 1
x
0
1
Yi
Let
i
{Yn, n >_ 1}
be i.i.d. matrices of the form
Prove that one can have
.
lim
IIYn...Y111
a.s.
limn Log IIYn
but
=0
.Y1 II
m>_
Exercise 2.6. With the notations of Lemma 2.3 we want to prove that n f+ dX < under the sole condition f o 61 = Ej(f) lim E
n
J
i=1 (i) For any K < 0
EK = {x
let
;
1E`T(f)(x)
> K}
.
Using that
n
EK e J prove that
lim n E i=1
(ii) Show that on {x ;
f
o
6' = lEJ(f) on
lE`T(f) (x)
EK
n
limn E f i=1
o
0i
24
(iii) Conclude.
11.3. Contraction properties
Consider a sequence order 2 with
of i.i.d. random matrices of
Y1,Y2....
= 1. We intend to show that under broad
Idet Y.
on zero
conditions, for any
x
IR2, IIYn...Ylxll -> W with
in
n -> - . The following deterministic result makes
probability one as
it clear that we have to look at the action of
Y1...Yn
on continuous
distributions on POR2). PROPOSITION 3.1.
with
Idet A n
Let
= 1
I
{An, n >- 1)
S
z
.
2 x 2
matrices
Suppose that there exists a continuous
.
distribution m on measure
be a sequence of
P(1R2) such that
A m converges weakly to a Dirac n
Then
lim IIAnII = lim IIAnII = and if
x
in
is a unit vector in
z
(5)
n-)
n->-
1R2 with direction
z
, then for any
IR2
IIAn4II lim n-wo
We first prove LEMMA 3.2.
A
Let
of order 2 and
= I<x,z>l
(6)
I I An *11
:
be a non zero (non necessarily invertible) matrix
m
be a continuous distribution on
P(1R2). Then the
equation f d(Am) =
dm(x)
I
J
valid for all bounded BoreZ functions, defines a probability measure Am
on
If Anm
P(IR2). (An)
is a sequence of non zero matrices which converges to
converges weakly to
Proof of the lemma is well defined. Since
Am
.
For any A # 0
x
in
1R2 such that
Ax # 0, Ax
there exists at most one direction
y
A,
25
such that if
is a vector with this direction,
y
is defined for all
except
x
continuous and therefore
Since
Ay = 0. Hence
m({y}) = 0
but
m
since
is
A -> A then for
can be defined. If
Am
outside a countable set
x
all
y
n , An x exists and converges to
D
m(D) = 0 , Anm -> Am.
Proof of the proposition
to some matrix A
Suppose first that
By the lemma (note that
.
converges
II-1 A
IIAn
n
IIAII = 1
A # 0)
so
,
Am = d
z
If
det A # 0
m
,
continuous. Hence
A
is the Dirac measure at
and is not
and
det A = 0
det An
0 = Idet Al = lim
1
I = lim i
n-,°°
II An II 2
11 An 112
proving (5).
A
Moreover the range of direction of this line is
z
canonical basis of IR2, Ae1 For any
x
is just a line and since Thus if
.
is the
Ae2 = ± IIAe2IIz
= ± IIAe1 llz and
,
IA*xII2 = 2 + 2
1A
Ae2>2
,
and
IIAe1 II2 + IIAe2II2 = 1 2
<x,z>2
= 1
e2>2
IIAe1 II2 + IIAe2II2) <x,z>2
IIAn*112 II=IA*xII2
n-
,
n
This is true for any convergent subsequence of
IIA
n II-1 An
,
so we have
proved the proposition.
*
So we are led to show that
*
Y1...Ynm
converges to a Dirac measure
at least for one continuous distribution m and let (i.e.
p 11
be the distribution of
* V = J MV d11(M) = v)
converges by Lemma 2.1
.
X1
.
If
on v
P(IR2). Put is
we already know that
Actually we have
X. = Y.
p-invariant X1...Xn V
26
PROPOSITION 3.3.
on
X1,X2,...
Let
be independent
and the same distribution
Idet XiI = 1
matrices with
2 x 2
Suppose that there exists
.
p-invariant distribution
P(1R2) a continuous
support of p
p
v .
Then, if the
is not contained in a compact subgroup of G1(2,]R),
there exists with probability one a direction X1(w)...Xn(w)v
converges weakly to
such that
Z(w)
dZ(w)
Moreover the distribution of Z is v and is the unique p-invariant continuous distribution on
Proof
P(]R2).
An = X1...Xn. From Lemma 2.1 we know that there exist
Let
Vw
probability measures
such that, for almost all
w
,
An(w)V -!L-> Vw w
A(w)MV
and
> vw
for
p-almost all
Fix such an w . For any limit point
M of
A(w)
IIA
n
(w)II-1 A (w) we have,
n
by Lemma 3.2
A(w)V = A(w)MV = vw for
M
p-almost all Notice that
H = {M E Gl(2,]R) ; MV = V
IIMnII -> W
case
and
CV = v
H
is a compact
such that C
.
In that
and
n1I)I Mn
=limJIM
v = Cv
sends all ]R2 onto a line and
C
in
converges to some matrix
IIMnII-1 Mn
IdetC( =lim Idet(JIM so
Idet MI = 1}
,
Mn
subgroup. For otherwise there would exist
1II2=0
n
must be a Dirac measure,
which is absurd.
is invertible then
A(w)
If
p(H) =
1
MV = v
for
which contradicts our assumption. So
matrix. If Vw = 6Z(C)
This proves that
Note that
A(w)
J
the distribution of
v
.
Since by Lemma 2.1
P(]R2)
on
f do = 1E { I f dvw} Z
is
and
A(w)v = vw
X1(w)...Xn(w)v W > 6Z(w)
does not depend on
Z
Borel function f
M
is a rank-one
is the direction of its range, then since
Z(w) .
p-almost all
V
and
v
]E {f (Z) } is unique.
for any
27
We do not say that there exists a unique 11-invariant
Remark 3.4.
P(1R2), but only that there is at most one such
distribution on
distribution which is continuous. Remark 3.5.
is contained in a compact group, the result
Supp(p)
If
does not hold. We have seen in this case that
M E Supp(p)
X1...XnV = V
so
My = V
for each
.
We now prove the following basic result
THEOREM 3.6.
Let
be a sequence of independent random
1}
{Yn , n 2-!
matrices of order 2 with the same distribution Suppose that
Idet Yij = 1
(a)
u
:
a.s.
(b) The support of
is not contained in a compact subgroup of
pi
G1(2,1R) . 11 -invariant distribution
(c) There exists a continuous P(1R2) where
pi*
Then (i) For each
is the distribution of in
x, y
on
Y1
P(1R2), with probability one Yn...Y1.y) = 0
lim
.
n(ii) If
x#0
, there exists
]E(Log+IIY1Ij)
0
lim a Log IIYn...Y1x1I = lim n Log nn-_ and if
v
is a p-invariant distribution on
(( Y= Proof
Set
distribution on set
such that for each
,
with
00
II
II
a.s.
=y
P(1R2)
dp(M)dv(x)
x
Sn = Yn...Y1
.
If
m
is a continuous
u -invariant
there exists by Proposition 3.3 a measurable
P(1R2) 1P (S10)
IIMxII
Log )
II
=
1
and a random direction
such that, for
wEl20 , *
lim Sw)m = By Proposition 3.1 if
x E 1R 2
,
then for all
Z
dZ(w)
is a unit vector with direction
w c 920
Z
and
28
lira IS.(W)
(7)
n-
'IS()xIl n
lim n>°°
and
For each fixed
= II
(8)
Sn(W) ll is the direction orthogonal to
u
x # 0, if
x
IP((w; = 0)) = 1P((w; Z(W) = u}) =m({u}) = 0 since
m
is continuous.
x # 0 , by (7) and (8)
So for each fixed
lira IIS.(w)xII
n-
ts n (W)xll
lim
and
+
a. s.
(9)
# 0
a.s.
(10)
n- Sn (W) II We thus obtain by (2), for
x,y c IR2-{0}
IIxII IIxII a(x.Y) lim 5(S n-wo
S
n
lim
n
n-5
= 0
a.s.
,
IISn(W)xII IISn(W)yII
Suppose now that IE (Log IIY1 II) < °° . (Note that Log IIY1 II >= 0 since Idet Y1I = 1 , see Exercise 1.5.1). We define an additive cocycle a on Gl(2,1R) x P(IR2) by IlYxll
G(Y,x) = Log
,
Y C Gl(2,1R) ,
x e IIt2-{0}
IIxII If
v
is a µ-invariant distribution on
P(1R2), then by Fubini's
theorem and (9) 1P® v-a.s.
lim a(Sn(w),x)
nSince
f a+(Y,x)dp(Y)dv(x)
Lemma 2.2
0
is in
0. Moreover, by Proposition 1.3.4, for some
(11)
limn Q(S.(w),x)
limn Log IIS.(w) x II = IIxII
((
: 0 x IP(1R2)->lR
IP 0 v - a.s. with ff 0(w,x)d1P(w)dv(x) = y
_
(w,x)
29
Fix an
such that (11) holds for almost all w. By (10)
x
1
lim In Log IISn(w)II = lim n Log IISn()XII = (w.x)
nSo
(w,x)
is the upper Lyapunov exponent
directly that
(one can also remark
y
0-algebra of the
depends only on the tail
1(w,x)
sequence of the independent random variables
Y1,Y2,...
and hence is
constant).
y # 0 , by (10)
Finally, for any fixed
lim n Log ISn(w)YII = lim
a.s.
Log IIsn(w) II = y n
proving the theorem.
One must note that the almost sure convergence statement in (i),
resp. (ii), is only valid for
x, y, resp. x, fixed. For instance, as
one can guess from the proof,
lim n Log IIYn(w)...Y1(w)xjj
to
y
x
if
is in the direction orthogonal to
is not equal
Z(w). The content of
Osseledec's theorem which we will prove in Chapter VI is that this limit is actually
-y
Let us show this on a typical, although
.
degenerate, example.
Example 3.7.
Suppose that
is. a sequence of i.i.d. bounded
(b.).
non constant random variables and set for 2
Yi
i ? 1
b1 .
=
0
1
n If
an =
2-1bi
E
and
Sn = Yn...Yt
then
i=1
2n[20
Sn = 2 -n/2 O
For
a. =
E
i=1
2-1b.
we have
1
2o
[bo Since
a
1
0 1
=
J
bi
[I
1
a l
2bo
has the same distribution as 2 bo +
the distribution of the direction of Since
__
a-
this shows that
(a ) is 11*-invariant. 4
is not constant it is easy to see that this latter
distribution is continuous. So we can apply the theorem. If we choose
30 -ate
x =
y = 2 Log 2
we find
.
But for
x =
( 1
Snx = 2 -n/2 and 11S
x112- Zn(1 + { E i=1
n
L2n(an_ao)1
bn+1}2
So
2i
2-n
0
Let ,
v
P(1R
2 )
are continuous.
be a 11-invariant distribution on
(x c P(1R); 2 v({x}) > a}
is finite. Thus if
P(1R2).For each v
is not
31
continuous there exists
such that
> 0
(i
v({-}) < (i for any x e P(]R2) F = {x E P(1R2); V({xl) _ (3} If
x
0
E F
and since
( = V({xo}) =
V=
p
rr
1{x
IJ
V({M
I
1
0
JJJ
and M
b
so
v({M
This shows that
GP
is contained in the closed subgroup
V({M
But
is finite and non empty.
(S
E F
for
11
almost all M.
H = {M E Gl(2,1R); If
L = {0} U {x E 1R2-{0}
dimensional subspaces of
;
for any
E F
x E F} , L
1R2 and
x E F}
is a finite union of one-
for any M
M(L) = L
in
in
,
G11
contradiction with the hypothesis. Therefore
is continuous.
v
To prove that each p -invariant distribution on continuous we verify that
P(]R2)
is
satisfies (iii) too. Suppose there exist
G * u
one-dimensional subspaces V11V2,...,Vr
such that for each i E{1,...,r}
and M in G * u M(Vi) = V. for some
j
If
E
Wi
(12)
is the orthogonal of
Vi
,
(12) gives
M (W.) = W. *
Therefore for each M E G1, _ {M (iii) doesn't hold for
J
1
u G
i=1
One can in fact weaken the hypothesis (iii) using
PROPOSITION 4.3.
If u
r
r
; M e G *} , M(U Wi) = U
and
W.
i=1
:
satisfies (i) and (ii) of Theorem 4.1, (iii)
is equivalent to (iii)' For any
x in
M E Gu}
P(]R2),
has more than two
elements.
Proof
(iii) _> (iii)'
suffices to verify that if
is obvious. To prove the converse it x1,...,xr
are distinct points of
P(1R2)
with {x1 ,. ..,xr}
,
for each M e Gu
32
M
r 5 2. Each
then
{x1,...,xr}
H = {M E GV
i.e.
subgroup of
of
induces a permutation (M)
GP
is a group homomorphism. The kernel
and
x.
;
and
GP
Gu/H
,
Vi _=
x1, x2, X3
H
is finite. By (ii)
consider three non zero vectors
r > 3
of
c
,
is a closed normal
...,r}
1
of
H
is not finite. If
x1, x2, x3
with directions
We can write
.
with a# 0
x3 = ax1 + x2 M
and for each
in
H
for some
,
(i # 0
,
,
0
A.
i = 1, 2, 3
Mx. = A.x.
.
This yields
A3ax1 + A313x2 = a3x3 = Mx3 = aMx1 + (MMx2 = ax1x1 + W, and
aX
= aa1, a3 =
3
Idet MI =
and each M
1
finite, so
r
0
and there exists a
v .
To check (ii) of Theorem 4.1 one can either use Remark 4.5 or,
more simply, notice that if
then for all
x
GP
in P (]R2) , and
is compact and M
in
MU *6X) =(6M*a)
A
is its Haar measure,
G1
6x=A*6X
This contradicts the hypothesis.
Suppose that (iii) doesn't hold. We then could find in
P(]R2)
such that, for any M
in
,xr} _ {x1>
Gu ,xr}
{xi.,x2,...,xr}
33 1
M(-
But this would imply that
r
r E
i=1
d_
x.
r
1
)
=
i
r
d
E
i=1
x.
, which is absurd.
i
11.5. Some simple examples
Given a sequence
{Yn, n >>-
1)
Gl(2,]R) with the same distribution
of independent random matrices in p
satisfying Furstenberg's
theorem one would like to compute the upper Lyapunov exponent or the p-invariant distribution on classical law of large numbers with respect to
y
y
and/
P(1R2). Unlike in the case of the cannot be written as an integral
only. Therefore this computation is most often
p
unfeasible.
y
we have described one of the only known models where
1.2
In
has a simple expression. We give here two related examples due to Furstenberg, valid in any dimension and a third one which appears in the study of Schrodinger operators.
We shall see more intricate examples in Chapter VI.
5.1. Rotation invariant distributions
Let SO(d) be the group of orthogonal
determinant one and mK mK(SO(d)) = 1.
measure on V
on
For any
d x d
the Haar measure on x
in
which satisfies
P(1Rd), and = mK * dx
P(7Rd) which does not depend on z
P(1Rd)
matrices with
SO(d)
.
is a probability
For any distribution
,
mK * V = J MK * dX d) (x) = J and dv(x) = md and
m
d
is the unique distribution on
P(1Rd) with satisfies for any
Mmd = and
It is clear that
(13)
M e SO(d)
.
is actually the image of the probability measure
and
on the unit sphere which gives the element of surface area under the canonical projection onto
P(1Rd).
As in Exercise 1.5.4 one proves that if
d(Ym dm
Let
d (x) =
IIYx11-d
x
is a unit vector
for each Y s.t. Idet YI =1.
Y1,Y2'... be independent matrices in
Gl(d,1R) with
34
distribution
Then
p
Suppose that the distribution of
.
(in other words
k c SO(d)
for all
m
d
k e SO(d)
ak * P * d
is a p-invariant distribution on
k-1
P(]Rd)
is the same
kYIk 1 = p
, Vk e SO(d)).
(Proof
:
for any
,
d)=(ak*p*a
U *and =P * (ak so, by (13)
k-1*ak)*md-(ak*u)*and
,
u *and = J (dk *p) *mddmK(k)=(mK*p)*md =mK * (11 *md) = md) In this case the distribution of
IIY1xll
-
does not depend on
x
on
the unit sphere. So this example is a particular case of the one introduced in 1.2.3
.
An other example where the invariant distribution can be calculated is the following. Suppose that there are two probability
measures a1
and
other words each
Gl (d, IR)
such that
A2
on
Y
can be written as
.
1
11 = X 1 * mK * A2 ,
A K B
1 1 1
independent matrices and the distribution of on
dSO(d)). Then
Al
where
.
K.
A1.,K1.,B
(in .
1
is the Haar measure
is the unique p-invariant distribution d on
* and
P(IR ),because for any probability measure
V
on
P(IR ), by (13)
(A1 *mK*A2) *v =]`1 * (MK * (a2 *v)) =A1 So if
V
is p-invariant
v = Al
* and
and this distribution is
invariant. We thus have Y _
IIYX II
( - IJ
Log
I
dp(Y) d(A1
IIx
md)(x)
II
_
AKB A'x
(
Log
dA1(A)da2(A') dmK(K)da2(B) dmd(x)
IIA'xll
Writing
Log
IIAKB A'xll = Log IIAKB A'xll
one finds
- Log IIA'xll + Log IIB A'xll
IIB A'xll
IIA'xll
are
IIB Axll ( d), 2(B) dA1(A) dmd(x) y = J Log
llxll One can also check this directly by noticing that exponent associated with the distribution of
K2 B2 A 1
is the
y K1
.
,
35
5.2. The Cauchy distribution
Define a bijective map
Tr
:
as follows
P(1R2) -> IR U {co} x
Xi
x1
x = (
if
x2
is non zero then
)
7r(x) =
L°° Gl(2,IR) on
One can transport the action of
for t E IR u {-}
1 (t)}
.
if
x2 # 0
if
x2 =0
2
P(IR2) by setting
If g = (a
b)
at+b ct+d
(14)
This is just a concrete realization of
DEFINITION 5.2.
The space
P(1R2).
endowed with. the action of
IR U {W}
G1(2,IR) defined by (14) will be caZZed the projective Zine.
The image under
of the probability measure m2
Tr
is the image of the Lebesgue measure on the circle 1
defined in 5.1 , under
?Tr 1[0 21 d0
Thus it is the Cauchy distribution, i.e. the probability
.
tan B measure on
IR defined by the density
1
Tr(1+t2) For
in
z = a + i(S
distribution of
C
with
where
(SU + a
U
(i
for the
Cz
> 0 , write
is a Cauchy-distributed random
variable. Using the fact that each matrix
g
Gl(2,IR) can be
in
written
`a
b
0
a
one sees that
In z > 0
,
g C
i
with
k
g =
=
and more generally, for any
Consider now the random matrix Y = (X variable with distribution
x E IR , z' E C x (1
So if
U
in
z
T
with
if g' i = z g Cz = gg' Ci = C gg ',i = C g.z
any
k E 0(2)
1
-1
Cz. Denote by
pz
.
O1)
where
X
is a random
its distribution. For
,
o)CZ, = (0
0 1)
(1
-1
o)CZ, = (0
is a random variable with distribution
1)G-1/Z' Cz,
,
independent of
36
X
,
the distribution of
Cz * C_
is
that is
1/z,
Cz_
1/z'*
We
can thus write *
uz and if
z = z' +
Cz, = C, 1/z' is a
Cz,
It is easily seen that by Theorem 4.1, if
pz invariant distribution on
]R U
satisfies Furstenberg's theorem. Therefore
}1z
is its upper Lyapunov exponent,
y(z)
IIM() II dpz(M) d Cz'(t)
Y(z) if Log I(1)f
2
Log (tx-1) +t
fJ l
if
2
1
2
dCz (x) dCz,(t)
+ t
If Lag{(tx-1)2+1}dc (x) dcz,(t) + + 2 f Log t2 dCz,(t) - 2 J Log(1+t2)dCz,(t)
But since
pz * Cz, = Cz,
,
for any bounded Borel
if f(ttt)dCz(x) dCz,(t) therefore
f
on
JR
f(t)dCz,(t)
Y(z) = f LogltldCz,(t).
By computation or using the fact that upper half plane and that
Cz,
Loglzl
is harmonic in the
is the Poisson kernel one obtains
y(z) = Loglz'I
11.6. Exercises
Remarks : After the exercises (6.4), (6.5) and (6.6)
it should be easy
for the reader to study the Lyapunov exponent of arbitrary distributions on p
*
Gl(2,1R). (6.5) shows that in Theorem 3.6 one cannot replace
11
by
in the hypothesis (c). Exercise (6.6) implies that it does not
suffice to suppose that
GP
does not leave any one-dimensional subspace
invariant to obtain Fustenberg's theorem 4.1
(6.1)
Let
{An, n >_ 11
.
One really needs (iii).
be a sequence of matrices in
that for some continuous distribution m
on
P(]R2), A
Gl(2,]R) such
nm
converges
37
weakly to
d_ . z
Show that this is actually true for all continuous
distributions.
Verify that working with
(6.2)
instead of
P(r12)
P(1R2), all the
results of this chapter are valid for matrices with complex entries (with the same proofs).
(6.3)
Prove that Theorems 4.1 and 4.4 are equivalent.
(6.4)
Let
be a probability measure on
pi
det M = 1}
Denote as usual by
.
Sl(2,IR) = {M E G1(2,IR);
the smallest closed subgroup of
GP
Gl(2,IR) which contains the support of
Prove that one of the
p
following holds is finite.
Gp
(i)
For any
(ii)
(iii)
QG
(iv)
Q GP
some
Q
has more than two points.
P(IR2)
,
a_1)
,
a#0
a-1)
,
a
a
e{(
c
Q-1
{(0
b e IR},for some QtGl(2,IR)
,
# 0} U t(0b
1
b
b # 0}
,
for
.
Let
(6.5)
in
x
Q-1
be independent matrices with distribution
{Yn, n _> 1}
p
of the form b.(w)
a.(M )
Y.(w)
=
1
1
0
a. Prove that if for some Q
1
x e IR, a x + b. = x 1
invertible matrix
a.(w) # 0
,
1
such that each
a.s., there exists an
1
Q Yi
Q-1
is diagonal (the
converse is clear).
b. Suppose that for any
1E(Log 1IY1II) < m (i) Assume
Q
,
Q Y. Q-1
is not a.s. diagonal and that
.
]E(LogIa1l)< 0
Prove that
.
u =
ai...an-ibn
E
n=1
exists with probability 1 and that the distribution is p-invariant. For
of
e1 =
prove that
((1))
v
of the direction
{a de + (1-a)v i
0
_ 0. Then there is
d
x
n + ' .
2H , A1...Anv
.
It is more convenient to work with the unit disk
D = {z e C
,
be defined by
Izl < 11
3D = {z e G
and
T(z) = y+ii
,
Izi = 1}
This map extends to .
3H
.
: H -> D
Let
T
and
T(8H) = 3D
39 -i)
i
and G'=(ab
Consider the map
G -> G'
Let
Q = (1 1
b)
a
a, b e C
:
lal2 - lbl2 = 1}.
,
defined by
geG
$(g) = Q gQ-1 g = (b
If for
a)
in
and
G'
D = D UaD we set
in
z
az + b
g'z = _
bz + a it is clear that
Therefore it suffices to show that if (i)' for some (ii)' if
x e DD
bn
lim inf n log 1IA1...An11 > 0
u
Ibn12
-
, A'l...A'v > 6x
DD
a 1
n
.
lim inf n Log lanl > 0
z e D
Since
and for some
large enough
n
Ia For any
Ian12
with
bn) an
(an 1
and
for any
x
,
is a continuous distribution on
v
Write
a > 0
A' = ip(Ai)
in
n
I
> en a
D
a u+b
b
n _n I
bnu+an
and if u e D
Iul
- -°l =
b
an
(16)
IanI2 Ian u+1I n
, 21
A'...A'.OI
_
x
,
-2na 1-p
la I2(1-p) n
lim
point of the sequence
p
< e
1
A'A' 0 - A' 1... n n n+1
,
such that
1
yields
is a Cauchy sequence. If
(zn)
Let
I
(17)
and m a weak limit
aD
b re an
Since
and some
is bounded we can
y
m i
and
aniibn1,
-> y
as
i ->
40
By (16), if
u E DD
and
yu +
# 0 , A..
1
x .
and m = d
\)({-y 1}) = 0, A1'...A' V -> d ni x proves (ii) and (ii)'
Therefore, since
which
x
.
COROLLARY 7.1.
If
are i.i.d. bounded matrices in
Y1,Y2,...
lim inf n
and if
there exists a random
0
Sl(2,7R)
x in
DH such
that, a. s. (i)
(ii)
Y1...Ynz -> x
for any
H
in
x
Y1...YnV -> 8x for any continuous distribution
We bring to the reader's notice that since of proof of Fustenberg's theorem, due to
on
v
aH
aH = P(1R2) , our method
[34], consits simply in trying
to find a converse of this corollary.
Consider independent matrices same distribution such that k t SO(2). Then for any g
g
Y1,Y2....
in
Sl(2,IR) with the
has a distribution independent of
k Y1k 1
Sl(2,IR) the sequence
in
Z0 =
is a Markov chain on the upper
g
half plane. To verify this first note that if we set, for any Borel set A
H
in
and
h
in
Sl(2,1R)
A)
=
(18)
J
then if J
because
h-1h' e SO(2)
.
depends
Therefore
really only on Let
Fn
be the 0-algebra generated by
a-algebra generated by
F IE
Zo,...,Zn
n{1A(Zn+l)}=
.
Y11 ...,Yn
One has for
and
Gn
the
h = g YI...Yn
=P(ZnA)
and
G G F IE n(lA(Zn+l))= IE n{1E n(1A(Zn+1))} = P(Zn'A)
41
So
Z
is a Markov chain with transition probability
n
P
By
.
Proposition 7.1 this chain converges a.s. to some point of the boundary aH
From 5.1 and 5.2 the distribution of this point is
.
CZo
.
An example of such a Markov chain is obtained by choosing, for a
on
mK
g E Sl(2,IR), p = mK * 8g * mK where
fixed
SO(2). Starting from
x E H
{z' E H ; d(z,z') =
circle
is the Haar measure
it goes to a point of the hyperbolic (see Exercise 7.3), this point
Log Ilgll}
being chosen with uniform distribution. Remark that if
g,h
is defined by (18), then for
P
in
Sl(2,IR),
P(gh.i,g.A) = f
So for any
in
z
H
,
P(z,A)
Conversely consider a transition probability (19). Define a distribution in
Sl(2,IR) and
B =
.
k, k' E SO(2)
p(kBk') = P(i, so
is a Borel set
B
g E B}
p(B) = P(i,B) If
H which satisfies
on
P
S1(2,IR) by, if
on
p
(19)
.
Ek * 11 * Ek, = u
P(i,B) = p(B)
and it is easily seen that if
i.i.d. matrices with distribution
P
chain with transition probability
P
,
.
Since
hypothesis required in example 1.2.3, if
lim
are
is a Markov
satisfies the
p
and
Zo = i
llxll = 1
Log IlYxlldp(Y) = IE {d(i,Z1)}
d(z,Zn) = Y =
n
{Yn, n >_ 11
Zn =
J
11
n where
d
is the hyperbolic metric introduced in Exercise 7.3.
For instance let
Pt
be the transition semigroup with 2
2
on H i.e. the LaplaceA = 2 y2 {a a2 + a } x ay Beltrami operator associated with the hyperbolic metric. Since each g
infinitesimal generator
of
Sl(2,IR) acts on
Pt
satisfies (19). It follows from the preceding results that if
H
as an isometry
A
commutes with
is the associated Markov process starting from
i
,
g
lim X t
t
and each Xt
Xexists ao
t-)m
with probability one and is in is obvious
:
the operator
A
DH = IR. But in that case this result is a scalar multiple of the ordinary
42
Laplacian
2 {a2Z +
a22}
1
ax
with
Bo = i
is the complex Brownian motion
Bt
for some clock
Xt = BT
,
So if
.
ay
T = inf {t > 0
Bt E a H}
z =
and
So
X. = BT where
.
a. Prove that
Exercise 7.3.
b. If
.
Tt
t
z'
i} = SO(2).
G ;
with
=
d(z',z) = Log Jig Prove that this defines a distance on
g,g' E Sl(2,IR)
1g, H
set
IrI
(the so-called hyperbolic
distance). c. Let
d(zn,zn+1) in
aH
with
Exercise 7.4.
{z n, n >_
C
lim zn = x
n-
Let
IE (Log IIY1Ib< w . IIYn(w)II < nr
1}
for
be a sequence in H
lim inf n d(i,zn) # 0
and
(Hint : write
Y1,Y2,...
such that for some
with
zn =
r > 0
C
Prove that there exists
be i.i.d. matrices in
Prove that for any n
.
G
An E G).
with
and almost all
w
large enough. Deduce the analogue of Corollary
7.2 under these hypotheses.
x
CHAPTER III
CONTRACTION PROPERTIES
Y1,Y2,...
We now consider i.i.d. invertible random matrices arbitrary order, say
d
.
of
This chapter is devoted to the study of the
basic almost sure properties of the products
Sn = Yn...Y1
. We shall
derive their salient feature, which is the contracting action of S n
on the set of directions. In particular we shall give (see 3.4, 4.3 and 6.1), following Guivarc'h and Raugi [34], a condition ensuring that
(a) For any x # 0 in
Iltd ,
lim
n
a. s.
Log II Snx II = Y
(b) There exists a unique invariant distribution on
(c) For any x, y in P(]Rd) lim
P(lRd)
a.s.
,
Log
0
m
where
d
n is the natural angular metric on
P(lRd).
(d) The two upper Lyapunov exponents associated with
(Sn)
are
distinct.
This result is basic for further developments.
A close look at the proof of these statements in Chapter II d = 2) leads us to introduce the following conditions. Let
11
(for
(resp. u
*
be the distribution of semigroup in
Y1
(resp. Y1)
and
Tu
be the smallest closed
Gl(d,]R) which contains the support of
require that
43
l1
.
One has to
44
(i) There exists a that
m(V) = 0
p -invariant distribution m on
for any proper projective subspace
(ii) There is a sequence
(Mn)
such that
Tu
in
P(IRd) such
V . nm
converges
to a Dirac measure.
In the first two sections we develop natural and tractable conditions on
T11
itself ensuring that (i) and (ii) hold. In Section 3
we generalize the method introduced in Chapter II. Namely we study the action of
S*
n
on measures on
P(IRd). We derive from this analysis a
key property from which we shall deduce the main results and we prove statements (a) and (b).
In Section 4 we give, under an irreducibility assumption, a necessary and sufficient condition ensuring that a.s. as
0
n -> - . For later use we shall be needing the stronger
statements (c) and (d). They are proved in Section 6.
(The
d
Lyapunov
exponents are introduced in Section 5).
We end this chapter with a complement section, which can be skipped at first reading, where we slightly improve Furstenberg's theorem. For the reader's convenience we only consider matrices with real entries. But all the given results are valid with essentially the same proofs for matrices in
Gl(d,C). Finally we must say that the method we
use is influenced by the paper [32] of Guivarc'h.
III.1. Contracting sets
We first define what we will call a polar decomposition of a matrix. Let
diag(al,...,ad)
denote
d x d
the
matrix A whose entries
A.
satisfy
ra.i 1,J We write
0(d)
DEFINITION 1.1.
LO
if if
i=j i#j
.
for the set of orthogonal matrices, 0(d) = {M E Gl(d,IR);
M*M = Id}
Let M be a matrix in
Gl(d,IR). We say that
.
M = K A U
is a "polar decomposition" of M if K and
U
are in
0(d)
and if
45
A = diag(a1,..,ad)
LEMMA 1.2.
with
2 a2 a ...
a1
>> ad > 0.
Any invertible matrix has a polar decomposition. Moreover are necessarily the square roots of the
a1 a a2 >_ ... a ad > 0
eigenvalues of M*M. Proof
Let
:
{f1., i
0 p
M in
T ,
let
be the square roots of the eigenvalues
is the index of T ,
sup{ai(M)-1 a1(M) ; M e T} = o0
if and only if Proof
:
i > p
Fix an integer
{Mn, n a 0}
in
T
i
,
2 5 i
IS d
,
and consider a sequence
such that
lim ai(Mn)_1 a1(Mn) =
sup{ai(M)-1
n-,
a1(M)
, M e T}
M = K A U is a polar decomposition of M we may always assume, If n n n n n passing to a subsequence, that Kn (resp. U ) converges to some n orthogonal matrix K (resp. U) and that 11A II-1A converges to some n n diag(a1,a2,...,ad) with a1 a a2 a ... > ad a 0 But orthogonal .
matrices are isometries, whence
IIMnII = IIAn1I
.
This implies that
46 lim
IIMnII-1 Mn = K diag(a1,...,ad)U
n-*By assumption the index of
is
T
so that
p
K diag(a1,...,ad)U
(since the rank of
ap # 0
ap+1 = 0
and
Max{i;, ai # 0}). The
is
conclusion comes thus from the relation
lim ai(Mn)-1 a (Mn) n-
=
c1
m on
A probability measure
1.5.
proper if for any hyperpZane
in
H
P(lRd) is said to be
IRd
x e H - {0}) = 0
m(x c P(1Rd)
The following result shows that the word "contracting" refers to measures on
the action on proper
Let
PROPOSITION 1.6.
be a contracting subset of Gl(d,IR). For any
T
proper probability measure (Mn
in
, n ? O}
as n-> m Proof
:
m on
P(1Rd) there exists a sequence
Mnm converges weakly to a Dirac measure
such that
T
P(lRd).
.
Let
(Mn)
be a sequence in
to a rank one matrix, say
For any
M .
lim Mn°x = But since m n-*_ is a hyperplane, we see that
such that
T
x e ]Rd such that
is proper and since
z
; Mx = 01
is the direction of the range of
M
for m-almost all
x
nMnm
Mx # 0 d
x
lim Mn - x = z so that
{x a lR
for m-almost all
lim M - x = n n-
It is now clear that if
converges
IIMnII-1 Mn
dz
converges weakly to
It is often difficult to compute the index of a given set. But we shall use the following.
Let T
PROPOSITION 1.7. T
is equal to
d
then
compact subgroup of Proof : Let
be a semigroup of Gl(d,IR). If the index of MI-11dM
{Idet
; M C T}
is contained in a
Gl(d,]R).
T' = {Idet
MI-1/dM
semigroup with index equal to
; M E T} d
.
.
It is clear that
Furthermore its closure
T' C
is a
is itself
47
a semigroup and it is easily seen that its index is also show that
is compact. If
C
is a sequence in
(M 1)
subsequence, denoted also by M
d
.
we can find a
C
for convenience, such that
n
Let us
IIM
n
II-1 M
n
converges to some matrix M . We have
Idet MI = lim Idet (JIM Mn1I)I = lim
IIMnII-d
n Since the index of shows that
IIM
n
the limit is in
is nonsingular and this relation
d , M
is
C
converges. Therefore M C
since
C
is in fact a compact group.
It is useful to notice that under the hypothesis of
Remark 1.8.
[
Proposition 1.7 there exists a scalar product ,
IRd such that
on
is orthogonal for this scalar product. This
Idet MI-11dM
is due the fact that if A
C
is a compact semigroup
C
G1(d,IR). By Theorem 9.16 of Hewitt and Ross [36] or by
Exercise 11.2.4 this implies that
for M E T
G1(d,IR) and
is closed. So every sequence in
C
contains a converging subsequence and contained in
converges in
n
II
C
is a compact subgroup of
G1(d,IR) and if
is its normalized Haar measure, then each element of
C
is
orthogonal for the scalar product defined by
[x,y] = J Exercise 1.9.
<Mx,My> da(M)
Suppose that a semigroup
.
T
Gl(d,IR) contains a
in
matrix M with a unique eigenvalue of maximal modulus, this eigenvalue being simple. Show that
T
is contracting. (Solution
Show that any real matrix of order
Exercise 1.10.
decomposition. Prove that any M
in
Gl(d,V
M* M , and where
Corollary IV.2.2).
has a polar
d
can be written as M=KAU
where A = diag(a1,...,ad), a1 ? a2 > ... ? ad > 0 roots of the eigenvalues of
:
K
being the square
and
U
are in
U(d) = (M e Gl(d,C) ; M*M = Id} Exercise 1.11.
Consider two polar decompositions
of a matrix M
in
G1(d,IR)
.
*
(i) Prove that
M = KAU , M = K'AU'
*
*
U el,...,U ed
are eigenvectors of
(ii) Show that there exists an orthogonal matrix
AB = BA ai
and
K = K'B
a the entries
,
Bi
U' = BU . If and
B.
B
A = diag(a1,...,ad) of
B
M M .
such that
prove that if
are zero. If a1> a2>...>ad
48
with
B = diag(E1,...,Ed)
verify that
ci = ±1
be a sequence in
Let (M ) n following are equivalent
Exercise 1.12.
Gl(d,IR). Show that the
(i) For some proper distribution m on to
.
P(7R
d )
Mnm
converges
d_ . z
(ii) For any proper distribution m to
on
d P(]R ), Mlm
converges
6- . z
(iii) The limit points of
range
1
IlMnll
Mn
are rank-one matrices with
lRz .
111.2. Strong irreducibility
Given a subset
DEFINITION 2.1. (i)
of
is irreducible, if there is no proper linear subspace
S
]Rd such that (ii)
of Gl(d,IR) we say that
S
M(V) = V
for any M in
S
V
.
is strongly irreducible if there does not exist a finite
S
family of proper linear subspaces of
1R1, V1,V2,...,Vk
such that
M(V1 U V2 ... U Vk) = V1 U V2 ... U Vk for any
M in
S.
For instance
S =
0
1)}
is irreducible but not strongly
irreducible.
NOTATION 2.2.
Given a probability measure
on
Gl(d,1R) we write
Tu
for the smallest closed semigroup (resp. subgroup) of
(resp. Gu)
G1(d,]R) which contains the support of p Remark that if matrices
p
M
subgroup of Supp p
V1,...,Vk
such that
.
are subspaces of
]Rd, then the set of
M(V1 U ... U Vk) = V1 U ... U Vk
is a closed
Gl(d,IR). Therefore
strongly irreducible Tu
strongly irreducible
Gu
strongly irreducible
The importance of strong irreducibility comes from the following
49
result. Recall that a probability measure
V
on
is said to be
P(]Rd)
p-invariant if V(A) = JJ A
for any Borel set
on
Gl(d,R)
be a probability measure on
p
such
is strongly irreducible. Then any }p-invariant distribution
GP
that
P(]Rd).
in
Let
PROPOSITION 2.3.
dV(x)
v
P(fld) is proper.
Proof
Let
:
be the set of
P(R)
k- dimensional subspaces of
]Rd.
Consider
R0 = inf it < d
for some
V(V) # 0
;
V
in
F(R)}
{x e P(]Rd); Y, E V-{0}})
(where
r = sup{V(V)
;
A = {V F P(20 )
and If
are in
V2
and
V1
and the definition of V1,V2'...,Vk
A
V c P(Ro)} ;
v(V) = r} V1 = V2
either
are distinct elements of
dim(V1 n V2) < to
or
v(V1 n V2) = 0
implies that
to
Therefore if
.
A
k
v(V1 U V2 U ... U Vk) = E v(Vi) = kr i=1
A must be finite.
so
On the other hand, for each
V
A
in
r = v(V) _ if 1V
dp(M)
V(M J
V(M
But
r
M -1-V
is in
V
A
in
.
A
.
,
Write
L
,
{M E Gl(d,]R); M(L) = L} p
.
strongly irreducible. Therefore
R0 = d
V
and
Prove that a closed subgroup
G
G/H
T11
.
If
is not
T11
is proper.
of
Gl(d,IR)
irreducible if and only if each closed normal subgroup that
is a closed
It thus contains
is a finite union of proper subspaces and
Exercise 2.4.
and
r
for the set of vectors which belong to some
From the foregoing
subgroup which contains the support of
to < d, L
M V(M
so that for p-almost all
H
is strongly of
G
such
is finite is irreducible.
Exercise 2.5.
Show that
Gu
11-invariant distribution on
is strongly irreducible if and only if each
P(]R) d is proper.
50
111.3. A key property
The main properties of product of random matrices will be consequences of the following.
THEOREM 3.1.
be independent random matrices in
X1,X2,...
Let
p , defined on (0, A, IP) .
GI(d,IR) with a common distribution suppose that
is strongly irreducible and write
T1
Then, if Mn = X1X2...Xn , for almost all p-dimensional subspace
Z(w)
Proof
:
V(w)} = 0
is orthogonal to
v
Mn(w)v
P(IRd) and
on
w
dZ(w)
for almost all
.
Vw
,
Henceforth we fix such an { JIMn(w)II-1 Mn(w), n > 1)
P(IRd). By Lemma
on
w
2-n-1 Un
A = E' n=o Consider a limit point
.
A(w)x # 0
If
n -> - ,
as
,
Gl(d,IR) w.r.t.
in
g
v
a probability measure
,
M(w)g V -> Vw
A(w)
of
when
,
goes to infinity along a subsequence. By Proposition 2.3
A(w)g x = 0}) = 0
N) ({-x E P(IRd) ;
A(w)g V = lim Mn(w)g V = Vw
course define
support of
A
being
X- almost all
for
by Proposition 2.3
A(w)g V
11.3.2). It is clear that
Consider now a sequence converges to a matrix is
p). For any
{gn ,
g
in
for all
A(w)g gn V = Vw
,
p
is closed. The
in
Tu
.
(1)
T such that IIgn II gn u (which exists, since the index of
ggn a Tu .
g
in
n a 11
T1
(We can of
.
see the analogous Lemma
this yields
,
of rank
h g
,
{g e G1(d,IR); A(w)g V = Vw}
T1i U {Id}
A(w)g V = Vw
Tu
.
converges weakly to
Consider a p-invariant distribution
such that, in the weak topology
and
For
.
is any non zero vector of V(w)
11.2.1 there exists, for almost all
n
V(w)
is contracting there exists a unique p-invariant
Tp
distribution where
matrix with range
p
x ,
IP {x
When
w there exists a
Rd such that any limit point of
is a rank
{ IIMn(w)II-1 Mn(w), n >_ 1}
any non zero vector
of
V(w)
We
for its index.
p
so that
51
A(w)gh x # 0, A(w)g
If
converges to
Therefore if
A(w)g
v({x E P(IRd) ; A(w)gh x = 0}) = 0 then
A(w)g gn v
2.3
v
converges weakly to
Since by Proposition
.
is proper, (2) yields that
for any
A(w)gh x = 0
If (3) were true for all by
A(w)g h v
(2)
g
in
in
x
the subspace
Tp ,
(3)
IRd
of ]Rd spanned
V
g E T11 , x c Im h} would be contained in the null space of
{g x ;
A(w) , and hence would be proper. But for any
in
g
Tu
and
g V = V
this would be in contradiction with the strong irreducibility of We thus can find
vw = lim A(w)gngv = A(w)ghv . Let By (4)
,
be the linear span of
V(w)
for the rank of
(Supp
h
is
p
(4)
{x E IRd - {0}
is contained in the range of
V(w)
x e Supp vw}.
;
A(w)gh and
dimV(w) _ 1) =
is at least
A(w)
dim V(w) = p
V(w)
is the range of each limit pointd of
For any non zero
Tu
for which (2) holds and then
g c Tp
set( H(x) = {y E P(IR )
,
<x,y> = 0).
v = J Vw d]P(w). Therefore
IP(x is orthogonal to
V(w)) = IP(Supp vw C H(x))
= IP(vw(H(x)) = 1 )
= v(H(x)) and this is equal to zero for Finally if V(w)
1
dZ(w) dIP(w) = v
is proper.
is contracting then
Tu
is reduced to a point
does not depend on
V
V , v
.
Z(w)
in
p = 1
by definition and
P(IRd). By construction
Since the distribution of
Z(w)
Z(w)
is
must be the unique p-invariant distribution.
The following consequence is of prime importance. We shall make an intensive use of it.
PROPOSITION 3.2.
Let
Y1,Y2,...
be i.i.d. random elements of G1(d,]R)
52
with common distribution
Sn = Yn...Y1
and
p
.
Consider a polar
decomposition S
Kn and
with a1(n)
>_
If
in
Un
.
is a strongly irreducible semigroup with index
T1
with
An = diag(a1(n),...,ad(n))
and
0(d)
>_ ad(n) > 0
...
= Kn An Un
n
{Un(w)e1,...,U*(w)ep}
(a) The subspace spanned by
to a p-dimensional subspace
V(w)
p , then
converges a.s.
.
(b) With probability one lim
ap+1(n)
n-
ap(n) = 0
and
> 0
Inf
n
1Sn11
(c) For any sequence
{xn, n >_ 1}
in
IISnII 1Rd which converges to a
non zero vector , 11 Sn (w) 11
a. s.
_1
We first prove
If
LEMMA 3.3.
is the distribution of
p
equal to the index of Proof
:
and
Tp
TP
V.) = U
so that the index of
Looking only at the
V
is.
TP
of
IR d
such that
for each M E T
V.
i=1
i
i=1
TP
(recall that rank M = rank M ). Suppose
Tu
that there exist proper subspaces Vi,V2,...,VR M( U
is
TP
is strongly irreducible if
TP = {M*, M E TW}
It is plain that
is equal to the index of
the index of
Y1
i
P
with maximal dimension we may suppose that
's i
they all have the same dimension. Clearly {MV1 ,
for any
M
in
implies that
Therefore
TP
.
... , MVR.} _ {V1 , ... , Vk}
If
is the subspace orthogonal to
W.
{M*W1,...,M'`WR} _ {W1,...,WR}
M( U i=1
W.) = U
W. i
i
for
M
in
Tu
M
for any and
in
V.
this
TP
is not strongly T11
i=1
irreducible.
By the lemma we can apply Theorem 3.1 to the
Proof of the proposition random matrices V(w)
X. = Yi
.
There exists, a.s., a p-dimensional subspace
which is the range of each limit point
M(w)
of {IISn(w)II
Sn(w),
53
n >-
11
.
Since
IISnII = a1 (n) S*
*
n
a (n)
a (n)
diag(1,
=U n IISn II
a1(n))Kn
...
a 12 (n)
1
Each component of this product lies in a compact set. If a2(w), ..
are limit points of
ad (w)
a (n)
Uu(w), K-(w)
Un(w), Kn(w),
a (n)
a2
(n) (w) ,
..
then
ad (n) (w) ,
,
M(w) = UU(w) diag(1, a 2(w),...,ad(w))K:(w)
is a limit point of rank
M(w) = p
S*(w). Therefore with probability one
IISn(w)II
and
a2(w) 2 ... >_ ap() > 0
i.e.
,
p+1 (w) _ ... = ad(w) = 0
{U*(w)e1,...,U*(w)ep}
The linear span of
.
is thus equal to the range of
M(w)
follows readily from the relations
, proving (a). Assertion (b)
ap(w) J 0
ap+1(w) = 0
and
.
In
order to verify (c) write
IIAn Un xnII2
I(sn xnII2 11S
11A
n II2
1
a (n)
II sn xn II 2 so that I
If I
Sn
xn -> x
II
p
2
> {aP(n)}
2
and
2 E1{a1(n)} 2 < xn, Un e1>
2
n
a.(n)
d
1
E1
*
<xn
, Un e i >
2
i
is the orthogonal projection of
y(w)
x
onto V(w),
then
lim inf n-
a (n)
I(s (w)xnll >
inf{p(w)} Ily(° n a1 (n)
II
n
IISn(W)II
By Theorem 3.1 IP (II y (w) II = 0) = IP (x
is orthogonal to V (w)) = 0
and (a) follows using (b) and Lemma 1.4
.
We now deduce from this proposition an easy, but important, corollary due to Furstenberg. Actually we shall see in Section 7 an analogous result when
COROLLARY 3.4.
matrices in
G11
is merely irreducible.
Consider a sequence
{Yn, n ? 1}
Gl(d,IR) with common distribution
(a)
1E(Log+IIY1II)
(b)
Gu
< oo
,
is strongly irreducible
of independent random p
.
We suppose that
54
Then (i) For any sequence
which converges to a non zero vector,
(xn)
a.s.
limn Log IIYn...Y1 xn11 = Y v
(ii),If
is a p-invariant distribution on
y=
11 Mx II
Log
P(lRd), then
_
dp(M) dv(x)
IIx 11
(iii) If moreover lE (Log+ IIY1111) ad (n) > 0
>_
be the square
By Proposition 3.2 if (xn) converges such that, a.s.
C(w)
Is n (w)xn II
0 < C (w)
1)
are independent
p . Suppose that
TP
is
strongly irreducible and contracting. Then x, y
(i) For any
in
P(IRd), with probability one
lim 6
Sn -Y) = 0
n-
(ii) There exists a random direction converges in probabiZity to
Z
Z
such that x c P(IRd)
, uniformly in
(iii) There is a unique p-invariant distribution
and 'or any continuous function f
on
P(IRd)
Sup 11E {f(S x)} - 1 f dvl -> 0 n
v
S *-x
n
.
on
P(IRd)
,
as
n -> c
x
Proof
: Let
a1(n), a2(n)
be the square roots of the two top
(x ) and (y ) are two sequences of unit n n n n vectors which converge in direction, Lemma 4.2 gives
eigenvalues of
S S
.
If
57
d(Snxn , Snyn)
a1(n) a2(n) II Snxn II
I Snyn II
a2(n) IISnII
I
IISnII
IISnII
I Snxn II
I I Snyn II
Therefore using Proposition 3.2 we have, a.s. d
n
lim n-
(5)
= 0
n 'yn)
which, in particular, proves (i). *
Let
be the common distribution of the random matrices X.= Y.
p
1
and
Rn = Xn...X1. I claim that as
n -> -
Sup IE X,y For if (6) does not hold we could find (yni)
(Xn
(6)
0 e > 0, a subsequence
n.
and
for which
ni'Yni)} > e
Without loss of generality we may assume that converge. By Lemma 3.3, (5) holds for
R
.
and (yn1.)
(xni )
and
n
lim d(Rn xn. , Rn yn ) = 0 i-)oo
1
1
1
1
which leads to a contradiction. We know (Theorem 3.1) that if distribution on
measure
P(lRd) , X1...Xnm
*
m
is the unique U -invariant
converges weakly to some Dirac
6Z . We have
IE {d(Sny,Z)} *
whence, since
_
Sup xEP(IRd)
]E
S
n n
has the same law as
R n
Sup 1E {6(R x,R y)} + 11E n
n
1
n
X,Y
The first term of the right hand side tends to
0
by (6), the second
one is equal to
lE {J 6(u,Z)d(X1...Xnm)(u)}
IE {J which converges to If
f
IE {6(Z,Z)} = 0. We immediately deduce
is a continuous function on
P(IRd), f
continuous. Therefore, using the foregoing
,
(ii).
is uniformly
1
58
Sup IE{If(S n x) - f(Z)I} -> 0
n ->
as
x
Thus
f dmI
Sup x
f(Z)I}
0
and
This implies that and
Sni(w) E Tu
Sni(w)m
for
,
m-almost all
y.
converges weakly to the Dirac measure at
Tu is contracting.
x
converges
,
IISni(w)II-1 Sni(w) converges to a matrix with range
This proves that
P(IRd),
w
z
z
59
Suppose that
is not strongly irreducible. Let
Tu
V1,...IVR
x E Vi} , where
V. n V. = o
lemma. Since
are the proper subspaces introduced in the ,
a = inf{d (Vi,V. ) If (i) holds we can find
Vi= { E P(]Rd
x c V1
,
i<j
1
and
V 2
,
S R} > 0 Sn(w) E Tp
such that
0. But (b) of the lemma implies that for
lim some
E Vi
i # j
and
E V.
for each positive integer
a > 0
.
Thus
n
It is usually difficult to know whether a given semigroup
Remark 4.6.
is contracting or not. The study of this problem remains to be done. Remark 4.7.
For arbitrary positive matrices, i.e. matrices whose
entries are
> 0
]Rd
a stronger statement holds. Let
,
= {x = (x1I...,xd) E ]Rd; x. > 0
,
called the
with the topology,
following property
:
if
and
0
there is
0(M)
0, b > 0, c > 0, d > 0
Sup dx Log p(ex) = ]R
(8)
show that one can suppose
T(p(u),p(v)) = T(p(1u),
p(x) = ax+b
A =Logy .
T(u,v)
v+2 c+1
if
and
v =
ad-bc # 0 ad
and that
60
c. Prove the inequality (8). d.
If
0 (x
,
x1
Y) = Log
define
Y1
- Log
.
I
y2
x2
matrices.
2 x 2
Show that (7) holds for positive
IR2
are in
x = (x1,x2), y = (yl,y2)
d
Exercise 4.9.
(case d
>_
H = {x = (xl,...,xd) E lRd, E x i= 1}
3). Let
i=1
and
B
x, y in H fl 1R+ we consider
the boundary of H fl IR+ . For
the two points
{ax + (1-a)y, a e IR} which lie on B,
on the line
a, b
and define
8(x where
[a,b
C ;
,
Y) = Log [a,b
x,y]
;
is the cross ratio (if
,
x,y] _ p'). For any
x = as + lib, y = A'a + }1'b
[a,b] = {aa+ (1-(x)b, 0 < a
L
A
of
defined by
1Y(u))
9'
p(L) C L
.
Prove that (7) holds with
(Use Exercise 4.8).
111.5. Lyapunov exponents
We shall now define the
d
Lyapunov exponents associated with
(Yn). They will in particular help improve Theorem 4.3. This is easily done if we work with the exterior powers
APlRd. For the reader's
convenience we recall the definition and simple properties of these spaces.
Let
E
be a
d-dimensional vector space with dual
E
.
For any
61
positive integer
E
u1,u2,...,uP
For
.
denote the vector space of alternating
APE
let
p
p-linear forms on
in
E,
in
f1, ... If
E
we
P
set
(u1Au2A...Au(f1,...,f) =det[{fi(uj)}i j] It is clear that
is an element of
u A ...A u
AP]Rd , we shall call
P
1
it a decomposable
.
p-vector.
The following facts are readily checked (see e.g. 7.7 of Loomis,
Sternberg [52]). LEMMA 5.1.
{e1,...,ed}
(i) If
{ei A ei A ... A ei
is a basis of E , then
is a basis of APE.
f i1 < i2 < ... < ip f d}
1
p
2
1
,
(ii) For u1,u2,...,uP in E, u1Au2A...Au is non zero if and P only if u1,u2,...,uP are linearly independent. (iii) Two independent
{ui
p-upZes
,
1
f i f p}
{vi,
and
1 _ ad(M) > 0
...
M*M. Then for any
p,
< p < d
1
IIApMII
G1(d,IR), and
,
= a1 (M) a2(M) ... ap(M)
.
Proof : Write a polar decomposition M = K A U with
A =
( M ) ,...
APIRd
,
ad(M))
,
Since
.
AP K and
1
and
But
IIAp MII = II (AP K) (ARA) (ARU) II = IIARAII .
(ApA)(ei
K, U e 0(d)
are isometries of
AP U
A ... A ei ) = ai (M)...a. (M)e. 1 p p
A ... A ei
p
1
whence
IIApAII = Sup{ai (M)...ai (M) LEMMA 5.4.
For
Log IIApMII
1
I
1 < i1 < i2< ... < ip Y2 > ... ? Yd , which justifies the y = y1
the upper Lyapunov exponent.
Idet(Yn...Y1)I = a1(n)a2(n)...ad(n)
,
therefore
Y1 + Y2 + ... + yd =lE(Log Idet Y1 I)
and all the exponents are finite if and only if
Logldet Y1I
is
integrable.
111.6. Comparison of the top Lyapunov exponents and Furstenberg's theorem
We now prove the main results of this chapter. Under the assumptions
IE (Log+ IIY1II) < -
and
Tu
strongly irreducible we
determine following Guivarc'h and Raugi the least
p >
1
such that
Y1
> yp. In particular we recover Furstenberg's theorem. As before
Y
? Y2 ?
1
following is
[34]
are the Lyapunov exponents associated with
? Yd
11
.
The
maybe the principal result of this book. It appeared in
.
Let
THEOREM 6.1. Log
be a probability measure on
is finite. We suppose
IIMII dp(M)
J
p
Tu
Gl(d,IR) such that
irreducible. Then
y I
if and only if Proof
is strongly irreducible and contracting.
Tp
Suppose that
:
>Y 2
is strongly irreducible and contracting (i.e.
T
U
the index of x
in
IRd
with law
,
p
Iu if ,
is one). From Proposition 3.2. we know that for any Sn = Yn...Y1, where
then
{Yn, n _>-
1}
are i.i.d. matrices
65
a2 (n)
IISx1I n Inf
n
and
# 0
IISnII
11A2SII n
= lim
Jim
n-
= 0
.
n- IISnII2
IISnII
This implies that for each non zero matrix M , a.s.
IISnMII2
lim
= +
(10)
2
n-
11A Sn II
We shall use Lemma 11.2.2 to deduce from (10) that
Y1 > Y2 Consider each matrix as an endomorphism. From the proof of the theorem
of Furstenberg and Kesten (Theorem 1.4.1) we know that there exists a }i-invariant distribution
of
IRd ,
B1
the set of norm-one endomorphisms
,
such that lim
n-_ n IP ® v1 -almost all
for
on
v1
there exists a
Log IISnMII
= Y1
(w,M). In the same way, working with
11- invariant distribution
v2
on
B2
(A2Yi)
the set of
,
norm-one endomorphisms of A2IRd, such that
lim n Log 11A 2
n
for 1P ® v2 -almost all Consider
B = B1
S. N II
= Y1 + Y2
(w,N)
X B2
.
It is a
G1(d,IR)-space if we put
2
(A Y)N
)
Y c G1(d,IR),
IIY MII
u1
{n
E i=1
on
B1
Let
a
(v® ® v2)
;
n ? 11
,
N E B
2.
1
By Lemma 1.3.5 it is clear that each limit point n
M E B
II (A 2Y)NII v
of
is a }1-invariant probability measure
X B2 :
Gl(d,IR) X B -> IR be the cocycle defined by a(Y
,
(M,N)) = Log
Since the projections of V
on
IIYMII2 IIA 2
and
B1
YNII B2
are
V1
and
v2
IP ®v a.s. lim n a(Sn, (M,N)) = lim n Log IISnMII - "m n Log IIA2SnNII n,_ n
n
= 2Y1 - (Y1+Y2)
= Y1 - Y2 On the other hand, by (10)
,
.
we have,
66
,IS"M112 lim a(S y1
> y2
i->
= -
lim Log
n-
n
nand
(M,N))
II A 2S
n
11
by Lemma 11.2.2.
Since a direct proof of the converse would be tedious, we make use of the fact that when
is irreducible, for any non zero
G11
x
in
1Rd
a.s.
lim n Log 1IYn...Y1x1I = Y1
nWe will show this in Proposition 7.2 that if
Y1 > y2
Taking it for granted we see
for each unit vectors
,
"m 6(S
n-'°°
.
This in turn implies that
IIA2Sn(w)11
S "m
n
n
x, y < 0
a.s.
n-'°° II So (w)x 11 1 S. (w)y 11
is strongly irreducible and contracting
Tu
(see Proposition 4.4).
More generally we have
PROPOSITION 6.2. that
Let
p
is finite. If
f Log+ IMII dp(M)
if its index is
p
be a probability measure on
is strongly irreducible and
Tu
then
Y1 = Y2 = ... = Yp > Yp+1
Proof
:
Gl(d,1R) such
It is obvious from Lemma 1.4 that
.
Y1 = Y2 =
Yp
By
Proposition 3.2 for any non zero matrix M , a.s. IIp+1
lim
II SnM 11Ap+1S.11
=
n-'°°
n-'m
From this relation one deduces that
_
II Sn 11
l im
a p+1 (n) using Lemma 11.2.2
yp > Yp+1
a proof similar to the proof of Theorem 6.1
.
In particular we obtain the following result of
(Furstenberg's theorem)
THEOREM 6.3.
on
If U
[211
I
(b) GP
Log
IIMII dp(M)
0)
67
Proof
By Proposition 1.7, the index
p < d
of
Tp
satisfies
Therefore by Proposition 6.2 (recall that if
GP
is strongly
:
irreducible then
p
is also strongly irreducible)
T11
yp > Yp+1
But
.
the definition of the Lyapunov exponents gives Y1 + Y2 +
'"
+ Yd = lim n Log
n-
IIAdSnl
lim n Log Idet S,J 0
y
If
Yi T(n);
H}.
are i.i.d.
T(1), T(2) - T(1),...,T(n+1) - T(n)
random variables. Notice that since finite states Markov chain
Y'(S
n
is the first time when the
T
returns to the identity, its
)
distribution has an exponentially decaying tail. It is also clear that are independent random matrices in
ST(1)' ST (2) S-1 )'...'ST(n)ST(n-1)
H with a common distribution property of
1E(Log
T
. Making use of the stopping time
A
one sees that if
,
1E (Log
IlYlll)
is finite then
is also finite.
11ST11)
We shall now prove that for each irreducible subgroup of
1
< m 0, S. E 0) j 0 , i.e. IP ( 3 n > 0 , ST (n) E 0) # 0 . This
TX = Tu fl H
Tufl H
M E TA). On the other hand there
latter condition is equivalent to t1,t2'...,tp
exist
in
such that
T,,
Gu = t11H U t21H U ... U tp1H for, since
F(T-1)
is a finite semigroup in the permutation group
is a subgroup and
Y'(T11
contains
Tu
tt,...,tp
Tu
in
Suppose that
m E {1,...,k}
.
Y'(TU1)
with
and it suffices to choose
Y'({tai, ..,tp1}) =
Ga(m) is not strongly irreducible for some
M
G.
in S
s
W.) = U 1
For any matrix A
in
This implies that
W1,...,Ws
of
Vm
,
M( U i=1
.
= T(GU)
Then we can find proper subspaces
such that, for any
t.A E H
must be a closed subgroup which
Y 1(T(TU1))
Therefore
.
P
W.
i=1
1
there exists some index
Tu
A E TX
t j
(since
j
TX = TW fl H)
such that and that
71
(t.A)(W1) C U
W.
A WiC U
Therefore
.
M1,...,Mq
and we can find
t.1 W.
i,J
i=1
such that for each A
Tu
in
A{M1 W1, ..., Mq W1}
{M1 W1, ..., Mq W1}
in contradiction with the minimal assumption on
r
.
Finally we prove M. Using (iv) and Corollary 3.4 we see that for
any
x . FV .
,
x. # 0
,
II = lim 1n Log lim !n Log IS(i) -c (n)
IIS(i)
y(i)
x II = i
T(n)
(14)
This leads to
lim
n and to, if
xi
Log II ST (n)
= Sup limn Log i
I
IIST(n) II = Sup
i
is the orthogonal projection of
lim n Log
x
y(1)
Vi
on
ST(n)xll =Siup lim n Log IIST(n)x.
y(i)
_Sup {i;x ¢O} i
If
, making use of the usual law of large numbers we have
]E (T) = a
so that by (14)
lim T(n n) = a
lim T1 Log
for any
II ST(n)x
n
n
For any positive integer
,
T(0n) < n
_ 1}
converges to a rank-one matrix.
IIAPMn II-1 APMn
In other words
is p-strongly irreducible (resp. p-contracting)
is a strongly irreducible (resp. contracting) subset
{APM ; M E T}
if
be an integer
of proper linear subspaces of APIRd such that, for all M
L ,
p
We say that
is p-strongly irreducible if there does not exist a finite
T
(i)
union
.
be a subset of G1(d,IR) and
T
Let
{1,...,d-1}
APIRd. We shall give in the next
of the set of linear automorphisms of
section some criteria with which one can check these properties on examples.
The following important result is taken from Guivarc'h and Raugi
[34]
.
THEOREM 1.2.
matrices of
Let
where
Sn = Yn...Y1
Y11Y2,...
Gl(d,IR) with common distribution
1E(Log+IlY111)
is finite and that, for some
p
are i.i.d. random
p . Suppose that
in
{1,...,p-1} ,
p-strongly irreducible and p-contracting. Then, if are the Lyapunov exponents associated with
yp # - °°
(i) When
is
T11
>
Y1 > Y2 >
Yd
p , the following hold
YP > YP +1
,
(ii) There exists a unique p-invariant distribution
vp
on
P(APIR )
and AP JJMx Yl
i=1
dp(M) dvp(x)
- II Log
Ilx M
(iii) For any non zero
in
x
lim n Log IIApSnxIl n Proof
:
Let
a matrix in entail that
q
11
APIRd,
p =
Yi
E
be the dimension of
G1(q,IR). If Tp(q)
p(q)
,
a.s.
i=1 APIRd. We may view each
APY.
i
as
is its distribution, the hypotheses
is a strongly irreducible and contracting subset of
G1(q,1R). Therefore, if
exponents associated with
a1
>_ a2
denote the two upper Lyapunov
{APY , n 24 1) n
we obtain by applying
79
Corollary 111.3.4, Theorem 111.4.3 and 111.6.1
to this sequence
that
(a) Al > A2
.
p(q)-invariant distribution
(b) There exists a unique
on
v P
P(APIRd) and ((
11
Mx
11
dp(q)(M) dvp(x)
= JI Log
A 1
IIx II x
(c) For any non zero lim
APIRd
in
a.s.
Log IApSn xli
n It is obvious that the theorem will follow from these assertions once we have proved that
p
)1 =
yi
E
a2 =
and
i=1
P-1 E
-
yi+ y +i P
i=1
A short way to verify this is to write a polar decomposition of S If
Sn = Kn An Un with a2 (n)
a1 (n)
Let
basis
{ei,
1
>_
... ? ad (n) > 0
5 i < d}
.
n 0(d) and An = diag(a1(n),...,ad(n)),
in
, we have
APSn = (ARKn)(APAn)(APUn)
denote as usual the canonical basis of
A ... A ei
{ei
i1 < ... < ip S d}
of
APIRd
,
.
IRd.In the
the
P
1
matrices
Kn, Un
APKn
APUn
and
are orthogonal and
APAn = diag{ai (n)ai (n)...ai (n) 1
1
;
:5 i1 < i2 < ... < ip < d}
2
Therefore the two top eigenvalues of
{a1(n)a2(n)...ap-1(n)ap()}
2
and
(APSn) APSn
are
{a1n)a2(n)...ap-1(n)ap+1(n)}
2
This yields that (see Proposition 111.5.6) P
a1 = limn Log a1()a2(n)...a
n-
P-1
(n)a (n) P
i=1
Yi
P-1
and
A2 = lim n Log a1(n)a2(n)...a
P-1 n)a P +1
n
(n) = i=1 E y. + y P+1
The proof is complete.
Exercise 1.6 provides an example where and 2-contracting and where
Tu
is strongly irreducible
y2 = Y3
As a consequence of this theorem let us prove
COROLLARY 1.3.
Suppose that
Tu
is p-strongly irreducible and
80
p-contracting for any
p ,
S
Then one can find a polar
.
= Kn An Un
n
and An = diag(a1(n),...,ad(n)) , such that
Kn, Un E 0(d)
with
-< d-1
Y2 > ... > yd
and
Consider any polar decomposition
:
write for any
APS
p
n
n
n
Unei A ... A Un
x
in
clear that one can find %*
E1
n
1
UnEn e1
n
converges for any
p
decomposition and to
111.5.6)
such that
.
E {1,-1}
E2 n
It is clear that U.
Y1 > Y2 > ... > Yd
converges to
An/n
{1,-1}
x
It is
.
such that 2n,* A Un(e1A e2)
such that
AP Un En...En(e1 A ... A e
converges to some
then
) < -
_
Un = diag(c1,...,Ed)Un
Let
.
Kn = Kn diag(En,...,En)
1E(Log+IIY1
we see that the Choose a unit vector
converges. With this procedure we find
E {1,-1}d
(En, En,..., Edn )
n
In the same way, since
x .
and
S
converges to
Unei
P(A2IRd) one can find
A e2)
A2Un s n E n (e
.
with values in
= E1(W)
of
n
If we apply (a) of
.
converges a.s.
e
actually converges to
converges a.s. in
n
n
)
{APYn, n > 1}
p such that the direction of
IR
= K An U
S
,L n
= (APK )(APA )(APU
Proposition 111.2.2 to the sequence directiond of
U n
Um . Moreover if
converges almost surely to some orthogonal matrix
Sn = Kn An Un .
p
and
is a polar
By Theorem 1.2, when
and (see Proposition
= diag(exp(y1),...,exp(yd)). By
A
definition )1/2n (S*
nn
= (U* A* K* K
n
n
n
U )1/2n = U*(A1/n)U
A
n
n
n
n
n
n
*
UU Ao U-
thus this sequence converges a.s. to
Exercise 1.4.
[56]. Let
1E(Log+1IY1I) < -
Y1,Y2,...
be i.i.d. matrices in
Gl(d,1R) with
.
K , K Y1 K
(i) Suppose that for any orthogonal matrix
and
Y1
have the same distribution. Show that Y, + ... + yp = 1E(LogljY1 e1 A Y1 e2 A ... A Y1 epll) (see I.2).
(ii) We suppose that the
d2
entries of
are independent
Y1
Gaussian variables with mean zero and variance one. Show that the hypothesis of (i) is satisfied. Fix an integer
p
,
1
GI(4,]R) by
:
Sl(2,1R), Y(A,B)
in
represents the linear endomorphism (i) Show that (ii) Let
closed in
'Y
Y(A,B)(M) = A M B-1
M(2,]R).
of
is a group homomorphism.
be the group
G
is the matrix which
'Y(Sl(2,]R)x S1(2,]R)). Show that
Gl(4,IR) and p-contracting for
p = 1, 2, 3
.
G
is
(One can use
Corollary 2.2 and the image of a pair of diagonal matrices). Prove that G
is strongly irreducible. (iii) Let
}i1,}12
be two probability measures on
A2 are the top Lyapunov exponents associated with
Sl(2,1R). If and
u1
that the Lyapunov exponents associated with the image under
Y`
are
p
u2 of
a1,
prove
p1 ® u2
X1X2' A11X2' x1x21, x11x21
(iv) Show that one can have
Tu = G
and
11X2
= A1A2
IV.2. Some examples
We first give some criteria which can be used to check that a set is p-contracting. We then study two important examples where all the Lyapunov exponents are distinct.
PROPOSITION 2.1. 1
< p < d (i)
,
T
Given a subset
T of Gl(d,IR) and an integer
the following statements are equivalent : is p-contracting .
p
82
(ii) There exists a sequence a2(n) >_ ...
a1 (n)
in
(Mn)
such that, if
T
are the square roots of the eigenvaZues
>_ ad(n) > 0
lim {ap(n)} ap+1(n) = 0. v(iii) There is a compact subset C of Gl(d,]R) such that
of MnMn , then
P E C , M e T Proof
:
Q e C}
,
{P M Q
is p-contracting.
If (i) holds one can find
such that IIAPMn1I-1APMn
T
in
(M 1)
converges to a rank-one matrix. Write a polar decomposition M =K A U
n
.
n n n
We have
APAn = diag(a1(n)a2(n)...ap(n), a1(n)a2(n)...ap-1(n)ap+1(n),...)
so that
119M nII
(n)
-1
ApAn = d iag(1
ap+1 a (n) P
,
indicate positive numbers smaller than point of
APK
and
APU
...), where the dots
ap(n) lap+1(n). Since each limit
is an orthogonal matrix,
n n converges to a rank-one matrix, i.e.
IIAPM II-1 A PA
n
n
lim ap(n)-la p+1(n) = 0
.
This proves that (i) implies (ii), and (i) follows from (ii) by the same reasoning. Remark that
ap(n)
IIAPMn112
',p+1(n) Thus, if
Qn
and
Pn
IIAp-1Mn II 119+'Mn II
are in a compact set, there exists
a > 0
such
that IIAPPnMnQn1I2
a (n)
p
1
a ap+1 (n) for each
n
a (n)
p
aP+1(n)
IIAP1PnMnQnII IIAp+1PnMnQnII
This implies that (ii) and (iii) are equivalent.
.
The following corollary is particularly useful (Guivarc'h has announced that the converse holds under an irreducibility assumption on
T). Recall that an eigenvalue
if
Ker(M-A Id)
say that
A
of a matrix M
is one-dimensional and equal to
is dominating if
COROLLARY 2.2.
A
A semigroup
IXI > IA'I
T
in
is called simple
Ker(M - X Id)2 . We shall
for any other eigenvalue V.
Gl(d,]R) which contains a matrix
with a simple dominating eigenvalue is contracting. If there exists in a matrix with
T
d
eigenvaZues of distinct moduli, then
p-contracting for each Proof : Let A
.
p
,
1
M be a matrix in
< p < d-1 T
is
with a simple dominating eigenvalue
It follows from the Jordan decomposition that
to a projection matrix with range
T
.
Ker(M - A Id)
.
11M2n11-1 M2n
converges
Since this space is one
83
dimensional, {Mn If the
, n >_
is contracting (see Proposition 2.1).
1}
d eigenvalues of
M have distinct moduli, then
a dominating simple eigenvalue for any is
p
,
1
< p < d-1
APM has
Therefore
.
T
p-contracting.
Tutubalin was the first to show that when p
has a density on
S1(d,]R) with respect to the Haar measure, all the Lyapunov exponents are distinct (see Sazonov and Tutubalin [63], Tutubalin E69]). He
assumed moreover that
is finite, and Raugi (see [59])
IE (Log2 IIY111)
dropped this condition. But, when
has a density, T11
p
is open in
Sl(d,]R). The following thus shows that we recover their results.
If
PROPOSITION 2.3.
contains an open set of
T11
S1(d,]R), then
is p-strongly irreducible and p-contracting for any Proof : Let
V
be an open set in
show that, for
n
S1(d,]R) contained in
large enough,
is contained in
.
p
Let
.
We will 0
1
18- all < e
then
,
b1 R(e) I
B
2
0
is still in
P V P-1
we can find
el,...,en
is a positive integer such that
n
If
.
such that
R(ei)R(82) ... R(en) = Id
.
lei-ail < E
This implies that
is in
If we carry on this procedure we find an integer m contains a diagonal matrix. We then perturb contains which has
P Vm P-1
P Vn P-1
.
Br
0
L
obtain an element of
2n E > 27r
and
such that
P VP-
slightly this matrix to eigenvalues of distinct
d
modulus.
Finally we prove that irreducible. Since which contains
V
Sl(d,]R) is
is easy to check that
Tu
,
or equivalently
G11
,
is p-strongly
is connected, the only subgroup of Sl(d,1R)
Sl(d,1R)
itself. Therefore
GP = Sl(d,]R) and it
Sl(d,]R) is p-strongly irreducible (use e.g.
Exercise 2.9).
M F- Tu}
{ldetMl-1/dM
It is of course sufficient to require that
Remark 2.4.
contains an open set of
{M c Gl(d,]R); Idet MI = 1)
to
obtain the conclusions of the proposition.
We have net in 11.5.1 and 11.7 probability measures which satisfy clear that that unless 1.3 hold for
EK * P * e K-1 = u
KTPK 1
= Tu
for any
is supported by
1i
p
for any
K
in
S0(d)
11 on .
S1(d,]R)
It is then
K . Therefore the following shows
S0(d)
the conclusions of Corollary
(see an interpretation of this in Section 4). Actually,
85
using some analysis one can deduce this result from the preceding proposition. But we prefer to prove it directly since the method can be used in other contexts.
PROPOSITION 2.5.
Let
be a semigroup in
T
Gl(d,]R). If there is a
matrix M such that (i)
IIMII-1 M
is not orthogonal
(ii)
For any
K e SO(d)
then 1
is in
1
T
is p-contracting and p-strongly irreducible for each
T
< p < d
Proof
, K M K
If
:
p
.
is the smallest semigroup which contains {K M K 1
T'
K E SO(d)}, KT'K 1
for any
= T'
K e S0(d)
and
Without loss of generality we will suppose that
,
is contained in
T'
T'-= T
T.
Using the
.
polar decomposition it is clear that one can write M = U
with
B U2 1
U11 U2
in
B = diag(bt,...lbd)
and
SO(d)
Assumption (i) implies that there exists for any
i # j
exists
.
We claim that for any
K e SO(d)
j
where
bi
are real numbers.
such that
K1,...,Kn
in
Ibjl > Ibil
such that
...
K(K1 B K1 1) (K2 B K21 )
(Kn B Kn1 ) c T
(2)
.
For proving this by induction it suffices to show that if K E SO(d)
there is some
U1K 1 P KU11 ET and
K' a SO(d)
such that
T
P C T e T
K'P K B K 1
and .
But
whence
U1 B U2 E T
K U2(U1 K 1 P KU11) (U1 B U2)U21 K 1= KU2 U1
is in
there
SO(d)
K-1
PK BK 1
.
Since
bjl >
Ibil
for
matrices, one can find
it is clear that for
i # j
enough and for a convenient
)
large
K1,...,KR-l among the permutation
of
A = (K1B K1
m
...
(Km B
)
such that
m A = diag(at,...,ad) By (2)
K A is in
,
and
for some
T
positive integer n Kn-1An-1 E T since
,
KnAn
{KnAn
K
is in
a1 > a2 > ... > ad > 0 SO(d)
in
.
.
Let us show that, for any
This is true for n = and if AK = K 1(KA)K e T then Kn 1An-1AK is in T
KnAn = K(Kn-1 An K)K 1
clear that
with
, n >_ 1)
T
.
1
,
is in
T
.
, hence
T
,
By Proposition 2.1 it is now is p-contracting for any
p
86
We now prove that say that the group K2A2
generated by
G
are in
KA
and
is p-strongly irreducible. This amounts to
T
T
U
is obvious that for each A 3K 3K3A c G
A2
so
is p-strongly irreducible. Since
T
G
.
Suppose there exist proper subspaces V1,...,Vr i E {1,...,r)
for any
and
(it
SO(d), U G U 1 = G). This implies that
in
is in
K3A = K(K2A K 1) E G
and
K2A K 1 E G
,
G
in
Y
,
AP1Rd such that
of
for some
APY(V.) = V.
j
J
iff
{1,...,r} (defined by
induces a bijection of
A2
Since
we can find an integer m
= V .)
(APA2)(Vi )
A2(i) = j
Am keeps
such that
J
fixed each of these points, that is, writing D
APD(V. )= V. Since
,
for any
is a linear homomorphism.
u c APIRd , v E Ad-PIRd. Show that
Prove that, for any M
in
Gl(d,IR) and any
u
in
(det M) 0 (AP M u) _ (Ad-P M*)-1 0(u)
APIRd .
87
Show that if then
T
is a p-strongly irreducible subgroup of
T
Exercise 2.7.
If
p
k_1 distribution of
be a probability measure on k
A e2 A ... A ep
11-invariant. Show that when
if
Let
only if APIRd
P(ApIRd) where
on
SO(d)
p
< p < d)
is a
ma
ma
is
SO(d)
,
Gl(d,IR). Show that,
is p-strongly irreducible if and
T
,
K
Prove that
P(APIRd).
be a connected subset of (1
.
is p-irreducible (i.e. there is no proper subspace
T
such that
p = 1
exists some
for any M
ApM(V) = V
T). Hint
in
:
V
for any
of p
is connected thus it suffices to consider the case p= 1.
{ApM , M E T}
When
T
is an integer
p
denote the
is not confined on
the only p-invariant distribution on Exercise 2.9.
Sl(d,1R) such that
ma
SO(d) . Let
in
random matrix whose law is the Haar measure on is
Gl(d,]R), prove
(d-p)-contracting.
is
for any
= p
Ek * p1 * E
E T}
1
Let
Exercise 2.8.
is a p-contracting subset of
T
T-1 = {M ; M
that
Gl(d,IR),
(d-p)-strongly irreducible.
is
use the fact that if such that for an
q
T
w
is not strongly irreducible there in
(1d)
is finite.
IV.3. The case of symplectic matrices
Let
Sp(d,IR) denote the group of order
i.e. matrices
M
in
G1(2d,]R)
symplectic matrices,
d
d E
2d
which satisfy
Mei A Mei+d
.
ei A ei+d
i=1 It is clear that the action of
Sp(d,]R) on
A2IR2d
is not irreducible
d
since the space spanned by lEl ei A ei+d
is invariant. Thus Proposition
1.2 is not suited for the study of Lyapunov exponents associated with random symplectic matrices. Since these matrices occur in many applications (see e.g. Chapter IV of Part B, and the references in Tutubalin [68]) it is worth proving a criterion adapted to this situation. Before stating it let us recall some basic facts. It is immediate to see that if
88
I
0
O
where
is the identity matrix of
I
Sp(d,1R)= {M E: G1(2d,3R)
transpose of
*
of M so is
(iii) There exist two orthogonal matrices
>= ad
Proof : We have thus
1
M- J(M )
eigenvalue of
K and U
in
such that
1
M = K A U
M*JM = J
so that
= J
MJM
and
If
M M
= -J
(iii) note that the above
such that, if a
{f1,...,fd, if1I... ifdI
are the eigenvalues of
1
is an
A
relation implies that we can find an orthonormal basis of form
J
(Jv)
To verify
.
.
But
.
v we have
M* J My =
M
1
M 1J 1(M*)-1 = J
= J , proving (i)
M with eigenvector
is an eigenvalue of
Sp(d,IR)
with
A = diag(ai,...,ad, a1'. ..,ad
M*(Jv) _ and
denotes the
Sp(d,IR) .
and a diagonal matrix a1 > a2 > ...
M*
(recall that
Sp(d,IR), then
is an eigenvalue
A
,
.
is in
(ii) If
Gl(d,]R),then
M* J M = J}
Given a matrix M in
LEMMA 3.1.
(i) M
M)
;
IR2d of the
a2 >- ...
greater than one, then
>_ as >
1
M*M f. = ai f. 1
and M M J fi = ai-2J fi
.
Let
denote the orthogonal matrix defined
U
by the relations U fi = ei
U J fi = -(ei+d)
,
.
We have * (U
* J U)fi =U* J e i = -Uei+d =J fi
(U* J U)(J f i )
and
-U* J ei+d
=
=
-U*ei
= -fi = J(J f i )
.
*
Therefore
a11,...,ad M = K A U .
U
J U = U and
Since
and
U
is symplectic. If
K = M U-1 A-'
K
,
A
and thus
A = diag(a1,...,ad
K
are symplectic and
is easily seen to be orthogonal, the proof is
complete.
From this lemma one readily deduces PROPOSITION 3.2.
If Y1 >_ Y2 > ... > Y2d
are the Lyapunov exponents
89
associated with symplectic random matrices, then Yd+i-1
p e {1,...,d}
For any
for
-Yi
be the subspace of
Lp
let
i = 1,...,d
.
A1IR2d
{Me1 A ... A Mep , M e Sp(d,IR)}, which is sometimes called
spanned by
the p-Lagrangian submanifold (it is known that u1 A u2 A ... A u is * p yi+l of Proposition 3.4 are valid for any p e {1,...,d}
a. We first prove that equivalent to the fact that
T11
is
Lp-strongly irreducible. This is
G
is
LP-strongly irreducible. But
u Sp(d,IR)
Gu = Sp(d,]R) (for
being connected,
Gu
contains an open
set by hypothesis). Applying Exercise 2.9 we see that we have to show V
that there is no proper subspace for any
M
A = diag(a1,...,ad, a1',...,ad1)
,
(Ap An)(V) C V
implies either that
V
for any
If
.
al > a2 > ... > ad > 1
the relation
,
n > 0
V , but then
is in
... A Mep , M e Sp(d,]R)} C V
is not proper, or that
V
(APM)(V) C V
such that
Lp
,
el A ... A ep
{MeI A and
of
Sp(d,IR). Suppose there exists such a
in
is in the orthogonal W
el A ... A e p
of
V
.
Butthen by (i) of Lemma 3.1, if <Ap M w ,
for any
M
in
w e W
v> = <w, AP M* v> = 0
Sp(d,]R)
and
is in W
Therefore Me A ... A M e
.
= W . This contradicts also the fact that
L
for
p
1
any M
v e V
and
V
is proper.
P
b. We now show that to
verify that
Tu
T11
p-contracting. For this it suffices
is
contains a matrix
M with
eigenvalues whose
2d
moduli are distinct (see Corollary 2.2). We first show that if U an open subset contained in
T11
,
then there is in
U
is
a matrix with
distinct, eventually complex, eigenvalues. We may of course suppose that -1
is not an eigenvalue of the matrices in
U
.
In that case the
92
mapping U
l
U -> M(2d,]R)
:
defined by 1(M) = (Id-M)(Id+M)-'
onto an open subset of
S = {M a M(2d,]R) , M*J + JM = 0}.
M
Since the eigenvectors of
and
to show that each open subset of
eigenvalues. For each M
in
let
S
are the same it suffices
*(M)
contains a matrix with distinct
S
be the resultant of the
P(M)
caracteristic polynomial associated with M
its derivative (see
and
M has a double eigenvalue iff
e.g. Lang [45j).
sends
is a polynomial on the vector space
S
P(M) = 0. But
It can be zero on an open set
.
if and only if it is the zero polynomial. This would imply that each matrix of
has a double eigenvalue, which is false since
S
diag(al,.... ad, -al,...,-ad)
is in
M
So we have found a matrix
-L ''d' a '" ''a 1
Suppose that
Let
of
U
with distinct eigenvalues are the real elements
A2k+1' " ''Xd
d
1
al,...,ad
for any
S
in
{fl,f2,...,fd' hl,...,hd}
be a basis of
2d
such that M(f2j+l + i f2j+2) = A2j+1(f2j+l + i f2j+2)
j = 0,...,k-1
.
M(h2j+l + i h2j+2) = X2j+l(h2j+l + i h2j+2)
j = 0,...,k-1
M fj = A. f.
j = 2k+1,...,d
M hj = A.1 hj
It is immediate to deduce from these relations that for any r,s, r# s ,
f*Jf r f*Jf r r s
and
If we replace by
hr
P(ei) = fi
Therefore if
M'
Sp(d,1R)
.
=0
,
a r
hr
f*Jh =0 r s f*Jh r r = ar # 0. this means that the matrix
P(ei+d)
1
0
< i,j k,k E
The spectrum of
be a bounded operator on a complex Banach space T
is
112
a(T) = {z c Outside
be the set of complex valued functions which are
F(T)
a(T)
bounded open set whose boundary is contained in
a(T)
of
U , we set
U
f(T)
Let
.
and if
and
f
T
be a
U
and
f E F(T)
is a finite union of
aU
C1
curves.
is analytic in a neighborhood
f
f au
It follows from the Cauchy integral theorem that on
.
is analytic.
R(z,T) = (zI-T)-1
analytic in some neighborhood of
If
has no bounded inverse}
z I - T
;
the resolvent
a(T)
Let
IE
depends only
f(T)
.
We shall use the facts that for
f, g
F(T)
in
(of + tg)(T) = af(T) + (ig(T) (fg)(T) = f(T) g(T)
f(a(T)) = o(f(T))
LEMMA 3.1.
Let
and
U1
boundary such that
the union of U1
N1
be two open sets in
U2
and such that
U1 fl U2 = 0
and
U2 .
I
a(T)
with
C1
is contained in
Define
R(z,T)dz
J
2rti
.
R(z,T)dz
N2
,
21
aU 1
Jau2
Then (i) N1 and N2
N1 + N2 = I (ii) For
are two projections such that N1N2 = N2N1
and
T(Ni(E))
i = 1,2,
spectrum of the restriction of
Proof and
Then
Let
:
U2
.
V2
and
V1
fi is in Ni
F(T) 2Tri
is contained in Ni(E)
to
T
is
Ni(E)
and the
a(T)fl Ui
.
be two non-intersecting neighborhoods of
i = 1,2,
For
N1T = TN1
.
let
fi
be the indicator function of
U1
V.
and
fauf1(z) R(z,T)dz = YT) .
Since
f1 + f2 =
Moreover
f
= f1
1
on a neighborhood of and
f1f2 = f2f1
For any complex a set ga(T) = TN1 + a N2
a(T)
,
f1(T) + f2(T) = I
so that (i) is clear.
ga(z) = zf1(z) + af2(z). The relation
yields that, if
T1
is the restriction of
T
to
113
N1 (E) , a(ga(T)) = 0(T1) U {a}
.
o(ga(T)) = ga(o(T)) _ (a(T) (1 U1) U {a}
But
{Q(T) n U1} U {a}
contained in
for any
a
0(T) U {a}
so that
is
This proves (ii).
.
We consider now an analytic family of bounded operators V
E E VI , where
{T(C)
every
a
is an open set in
V we can find
e
This means that for
.
CC
and a sequence
E > 0
(A
n
)
of bounded
operators such that
(1)
En IAn1I
0
E
N
IIT(0)n-NII
0
such
c{1 +2p).
notice that since
This implies that restriction of
0.
n e IN
Il
Proof
T( Q .
T(0) Ker N C Ker N
T(0)
to
N(E)
lim T(0)n = N , NT(0) = T(0)N = N. and
T(0)N(E) C N(E). The
is the identity and on
Ker N
T(0) = T(0) - N . Therefore the spectral radius of the restriction of T(0)
to
Ker N
is
lim II(T(0) - N)n[IIIn = lim IIT(0)n - NllhIn = P
.
114
This implies that
a(T(0)) G B(O,p) U {1}
.
is a small circle around 1, then for
in Ker N ,
x
Moreover if
y
R(z, T(O))x
is analytic in the disk whose boundary is
y , so that by
the Cauchy Theorem
2ii For
x
in
N(E)
, T(O))x dz = 0
R(z
J
Y
R(z, T(O))x = (z-1)-1x
,
and
R(z, T(0))x dz = x
1
2i fl(
Whence N = 27ri
R(z, T(0))dz
J
Y
2) When
is in
E
If
This shows in particular that 3) Let
D1
be the disk with center
disk with center B(0,p)
and
0
{1}
and radius
, z
6f U}
is contained in
132-
I
U
,
and radius
13p
, D2
the
an open set which contains
is contained in
D1 U D2
,
and
.
IT() - T(0)II
We shall suppose that
D1
U
such that
M = Sup{ IR(z,O) II o(T(E))
is analytic.
U
.
Let
y1
< M . In that case, by 2),
(resp. y2)
be the boundary of
(resp. D2). We define
N(E) = 27i J
R(z,E)dz Y1
Using 1), 2) and the preceding lemma we see that analytic, that
N(0) = N
and that
Id - N(Q = 271ri J
N(E)
R(z,t;)dz Y2
E -> N(C)
is
is a projection. Moreover
115
Let us show that if
N(E)x =x and
Suppose that
then the rank of N(Q
IN(S) - N(0)11< 1
one there exists some
N(E)y = y
in
A
I
Since the rank of
.
N(0)) (x- ay) II IIx - ay II = IN(Q (x - ay) II = x = ay when IIN(E) - N(0)11< 1 .
is
N(O)
N(O)(x - ay) = 0
such that
is one.
.
We have
=
c > 0 such that for IEI < E , and E -> N(C) is analytic.
If
A(E) C E
x' E E'
and
be such that
0
then the rank of
ICI < E
II = 1, Nx
IIx0
0
= x
0
and
is one and there
N(2;)
such that
The formula
0
choosing
0
is analytic. We find the wanted decomposition
shows that
Q(C) = T(E)(Id - N(Q).
4) To prove (iv) first note that, for since it is an eigenvalue of T(E) 2+
continuous, thus
Now for each
a() n
in
>
.
A(0) =
But
is in
IEI < E, A(1;)
and
1
A
U
is
3P IN
T(,)n (Id-N(Q) = 21i
1
zn R(z, )dz .
Y2
Therefore
dP dEp
Qn(C) -
IldEd'ap
and
for c = 27f Sup {
11-!L
d
P
1
2711
II
Y2
zn d P R(z,t;)dz dip
< c {1 32P}n
z C Y2
II
IEI < E}
The proof of the lemma is complete.
Remark 3.3.
It is easy to modify the proof in order to show that if,
for some m e IN, the family
{T(Q, E c V}
is merely of class
Cm
then an analogous result holds. (In the statement one has to replace analytic by
Cm
and the integer
p
in (iv) has to verify
0 0
is an
small enough,
IL(a) near zero. Namely,
analytic family of bounded operators on
PROPOSITION 4.1.
studied in Section 2.
P
is the operator
T(0)
a with
J
0 < a < 2 T
there exists
such that
q > 0
analytic family of bounded operators on
{T(z),
IzI < p}
is an
IL(a).
The proof is tedious but straightforward. It will require the following lemma in which directions x
LEMMA 4.2.
y
and
For
that, for any
and
z e M
M in
denote two unit vectors with
0 < a < 1
, there exists
c1, c2 > 0
IMxMI
-
__
Log IMYII
ci exp{[t4a)IRe z1+2a4 1r(M)
I
c
6(x,y)a
x,ycP(IRd)
Proof of the lemma
_ ez Log I1 MYI I,
6(x,Y)
Log
(ii) _ Sup
I Mx I I
-a
x,ytP(IRd)
a) First notice that for
6(x,y)2=1-<x,y>2=IIx-Y Replacing, if needed,
y by -y
2
f(M) e
2ak(M)
IIx1I = Ilyll = 2 x
112{1-114Y
we may suppose that
Ix-YII
DF
x
satisfies
for
1
}.
this case 1-_
such
Gl(d,IR)
ez Log
Sup
y
and
x
.
h c IRd
-
In
117
and
0
n E IN
aP II
1+2P
Qn(z) II < c
dzp
The following will be used for studying the speed of convergence in the central limit theorem. Under the hypotheses of the theorem, if IL
is the function identically equal to one,
COROLLARY 4.4.
Suppose that for some
X"(0) = a2 > 0 . Then there exist that for
n E ]N and
a e (0,ao], A'(0) = 0
13 > 0
,
c > 0
0 < r
0
(iii) There exists
Log IISnxII-nY Sup 11P (
< t) -
of
IIxII=1
of
A
ezLog
of Theorem 4.3, and
S.-x)
converge in
J
T(z)
t c lit,
t exp - u2 Z du I < C- .
-
is defined by
JIM. 11
P(1Rd) and
is a complex valued function on
f
notation
1
f2Tr
We first prove a lemma. Recall that
T(z) f(x) =
,
Yn-
.
IIxII = 1
.
,
For each positive integer n and unit vector
LEMMA 5.2.
x
Tn(z) f(x) = 1E{ezLog If
A(z)
is the eigenvaZue of T(z)
(6)
a'(0)
0 < a < ao
In particular
Y)21
a = 0
converges as
- n a2
and for
and
IzI < e)
a2= X"(0)- y2,
n -> - to a finite limit.
if and only if the sequence
{]E(Log 1ISn xII- n Y)2'
is bounded.
n >= 11 Proof
y
is the upper, Lyapunov exponent
1E { (Log II Sn x II - n
IL(a) with maximum
acting on
modulus which appears in Theorem 4.3 (for then
Using the
for the first two derivatives
A', A"
:
It is worth noting that (6) is only a consequence of the cocycle
property of
a(M,x) = Log
IIMXII
We proceed by induction. If Tn-1(z)f(x)
z Q(S = ]E {e
n-1
,x)
then
T"(z)f(x) = Tn-1(z){T(z)f}(x) z 6($ = 1E {e
,X) n-1
_ (T(z)f)(Sn_1x)}
123
= 1E {e
za(S n-1 ,x) zct(Yn ,S n-1 x) e
z U(Sn',x)
= IE {e Sn = Yn Sn-1
since
and
f (Y ( S
n
n-1 'X)) }
f(Sn-x) l
U(Sn,x) = U(Yn, Sn-1'x) + U(Sn-1,x)
This proves (6). We now prove the rest of the lemma. If
t
is a
small real number we can write by Theorem 4.2 T(t) = X(t)N(t) + Q(t) Tn(t) _ Xn(t)N(t) + Qn(t)
and
But for
IIx1I =
1
tLog IISnxII
Tn (t) 11(x)
IE (e We thus have
tLog Ilsnxll)
(dt n(t))N(t) IL x)+an(t)
dt1E(e
dt.(N(t) IL (x))
+dt {Qn(t) 1I(x)} so that, using Theorem 4.2
,
tLo g 1 d Y = lim n1E(Log IISnxII) = lim nl dt1E(e
IISn xII
)(0) = A'(0).
Now
a2 (e
tLog IIS xII -
2
dt
Notice
_
{e
2
e n") (0) = n(X '(0)-Y2)N(0) IL (x) + a 2{N (t) 11(x)}(0) dt 2 + 2 {e nt Y Q"(t) Il6o)(0) dt
d
that, by (v) of Theorem 4.2
nt Y
dt2
n
,
Qn(t) IL(X)1(0) = n2y2 Qn(0) 11(x) - 2n y dt {Qn(t) 11(x) } (0)
_
2
+ d 2 {Qn(t) Il(x)}(0) dt
converges to zero. Therefore, since
N(O) 1 = IL
_ IE{(Log IISnxII -nY)2} - n (U(0) - Y2) - a2 {N(t) 11(x))(0) dt
converges to zero. This proves the lemma.
We postpone the proof of the following lemma until section 8 (cf. Corollary 8.6).
LEMMA 5.3.
Under the hypothesesof Theorem 5.1 the sequence
IE{(Log IISnxII -ny)2) is not bounded.
1 24
Proof of Theorem 5.1.
:
Although (i) follows readily from (iii) we
shall prove it directly. By Theorem 4.3 and Lemma 5.2, for
small enough,
1E{e
f
in 1L(a)
and
II x I
(Log IISnxII-nY)f(Snx)}=e-ir ty{a(
whenever
is real and
t
But
lim
e-if t y
a > 0
i = 1 , we have )0N(
it )f(x)+Q°( -
)f(x)} VIE
sufficiently large.
n
2 2 )n _ exp - t2 a
X(
n-*
a'(0) = y
since
a"(0) = a2 + y2
and
a # 0 since, by Lemma 5.3
,
by Lemma 5.2 . Notice that
IE (Log II S. x I I
-n
Y)2
is not bounded. On the
other hand
lim N()f(x) = N(0)f(x) =
f dv J
n-
lim Qn(7) f(x) = 0
and
n4 (use (v) of Theorem 4.2 with
p = 0). Therefore for any
i f(Log IISnxIl -ny)
_
lim1E(e
in IL(a)
f
t 2 a2
f dv
n--
Taking into account that each continuous function on a uniform limit of a sequence of elements of IL(a)
P(3Rd)
is
we immediately see
that
(7n (Log II Sn x II - ny) , Sn-x) converges in distribution to
N(0, a2) ® v
.
To prove (ii) it suffices to use (i) and the fact that for any
x
0
,
is a. s. bounded (see (a) of Proposition
Log IIS n x lI - Log II SnI I
111.2.2). y = 0
We now prove (iii). We may suppose a'(0) = 0
and
X"(0) = a2 > 0
.
(see Lemma XVI.3.2 of Feller [18])
all A > 0
and real
M > 0
there exists
:
t
1
In this case
.
Recall the Berry-Esseen inequality
VF_ 1
(t
J
u2
e
2
1
gn(u)-e
A
duI _T J -
U
A
where
¢n(u) =1E{exp a,r Log IISnxII} = Tn( a/n=) 1(x)
By Corollary 4.4 there are when
Jul
< S
such that for
2
> 0
,
c > 0
and
2
Idu+AM
.
0 < r < 1
such that
125
Sup IIx1I=1 Choosing
0 1
such that for any
(Log II S n
ny)
x # 0
in
]Rd
converge in distribution
.
We shall be needing the following algebraic lemma which is proved in Chevalley, (45, ch. IV of [13]).
LEMMA 5.5. 1
_ 11 is almost surely bounded for each x # 0 in 1Rd . Whence An- {Log I Sn x 11 - n Y} and {Log II Sn I I -n Y} converge to the same distribution (if one of them does converge). Now let p denote the index of Tu . By Lemmas 111.1.4 and 111.5.3 the sequence {Log I I Sn I I p - Log II AP Sn 111 , n > 11 is also bounded so that 7 {Log II Sn 11 - n y} and P {Log II AP Sn I I - n y p} have the same limit
{Log 11S,111 - Log I Sn x II
I
distribution. I t is therefore enough to show that
{Log I I AP Sn II - n p Y}
converges in distribution to some gaussian law N(O,b2), b > 0 {IdetMI-11d
whenever
M ;
is not in a compact group, i.e. when
M F_ G
p < d (see Proposition 111.7.1).
y = y1 ? Y2 = ... > Yd
Let
with
p
.
be the Lyapunov exponents associated
We know (see Proposition 111.6.2) that
but that
Yp+l < yp associated with (APS ) n
X
Y i=1
for and
Al
1 X2
Consider now the direct-sum decomposition
AP]Rd=W1 ®W2 ® ... ®Wk given by the preceding lemma, applied to
G = Gu
.
It is easy to see
that the set of all Lyapunov exponents associated with
(APS )
union of the sets of all Lyapunov exponents associated
with the
restriction of choose
W1
(APSn)
among the
restriction of
to each
Wi's
APSn to Wl
W2 9 ... 0 Wk , then
W.
,
Bn
is the
1 < i < k . Therefore we may
in such a way that, if and
n
An
the restriction of
is the
APSn to
127
lim -1 nLogIIAn1I n4co
lim n Log II Bn II < A
and
2
n-*
A2 < X1
Since
and
prove that N(O,b2)
.
I
I Ap Sn I
IA
n II
, J I Bn II)
, we are done if we
converges in distribution to some
{Log IIAnII- n A1}
Let us check that we may apply Theorem 5.1 to the sequence r = dim W1
(An). If
= Sup(
I
random matrices on
we may view
r > 1
first show that sends all of
Ap]Rd
decomposable
p-vector
r = 1
If
.
n
as a product of
An
i.i.d.
with a common distribution, say
Gl(r,]R)
each limit point of
into the line
p
. We
IIAPSnII-1 APSn
. Therefore there is a
W1
v1 A ... A v
which spans
W1
But
.
p
(AP M) (W1) = W1
for any
M
1Rd spanned by
v1..... vp
,
Gu
in
,
M(V) = V
hence if for all
contradiction with the irreducibility of
G11
is the subspace of
V
M
in
G11
in
,
. Knowing that
there are at least two Lyapunov exponents associated with largest one is
with
(APSn)
,
.
The
and the next one, being also an exponent associated
Al
is not bigger than
exponents associated with irreducible on
r > 2 p
A2
.
Therefore the two top Tp
are distinct. Since
p
is clearly
by construction (see (a) of the above lemma) we
W1
know, by the converse statement in Theorem 111.6.1, that
Tp
is
actually a contracting and strongly irreducible subsemigroup of Gl(r,1K) . We may thus apply Theorem 5.1 to the probability measure and
with
(Log IIAnII -n p y)
b # 0
Remark 5.6.
.
converges in distribution to some
This completes the proof of the theorem.
Suppose that
compact subgroup of
{Idet
MI-1/dM,
is contained in a
M E Gp}
. We know (see Remark 111.1.8) that there
Gl(d,]R)
exists a scalar product on ]Rd
for which all these matrices are
orthogonal. If we work with the associated norm, then for any unit
vector x in ]Rd n
Log I I Sn I I= Log IISnxII = a Log I det sn I
aE i=1
Log Idet Yi I
is a sum of i.i.d. real random variables. Therefore, if
Y = ]E{Log I det YiI } ,
then {
converge to N(O,a2)
where
Note that subgroup.
p
N(O,b2)
a = 0
vin
Log I I Sn I I - n y} and a2
if and only if
rn {Log I I Sn x1 1 - n y)
is the variance of Log Idet YiI e-Y
Y1
is contained in a compact
128 r
In these theorems the assumption that
J
exp T R(M) dpi(M)
is finite
-"
is unpleasant. The condition
seems more natural.
j i(M)2 d}i(M)
0
,
has a
.
Sketch of proof : We define by (2) a family of operators
acting on the space e of continuous functions on
{T(it), tclR}
P(1Rd). By
Proposition 2.6 (see also Exercise 2.9) we have a decomposition analogous to the one given in Theorem 4.3
For
.
t
small enough,
T(it) = X(it)N(it) + Q(it) where X , N
and
have the same properties as in this theorem except
Q
that they are only of class
.
To verify this, note that if
t -> T(it)
is finite, then
J'i2(M) dp(M)
remark 3.3
C2
is of class
C2
and use
The proposition is thus proved as was (i) of Theorem 5.1
.
.
Making use of the arguments developed for proving Lemma 111.7.1 it is not difficult to show the following.
PROPOSITION 5.8.
random matrices Theorem 5.4
.
Assume that the common distribution
If
p
of the i.i.d.
satisfies the assumptions (a) and (c) of
(Yn)
GP
is merely irreducible
1Rd
=
then there exist a direct
sum decomposition
and a real number
x
in Vi ,
Vn-
a > 0
V
V
such that for any
Log { II S. x l I - n y)
1
< i < k and any nonzero
converges in distribution to
N (0, a2) .
But it is worthwhile to remark that it may happen that for some
nonzero
x
in IItd
,
1 {Log IISnx II -ny}
does not converge to N(O,a2).
Consider for instance the case where for some sequence i.i.d. random vectors of ]R2- (0,0)
,
(an,bn)
of
1 29
a n
or
a-1
0
b-1 n
0
0
n
=
Y
b
n
0
n
b11
0
If 1P(Y1 = (b
then
0 )) # 0
is irreducible. It follows from
G11
1
Proposition 111.7.4
T Log
y = 0
that
For
.
x = (i)
and
IISn xII ?1
so that this latter sequence cannot converge to a
S. x11 > 0 ,
centered gaussian distribution. Notice that the same remark holds for
7 Log IISn 11
.
In this particular case
N(0,a2), if
some
x = (0) or (I
Log I I Sn x II
. And
when
a # 0
does converge to Y1
is not a.s.
orthogonal.
In the following exercise we propose to prove, using the arguments developed in this chapter, the central limit theorem for positive matrices due to Furstenberg and Kesten [25]
Exercise 5.9.
of matrices
Let
c > 0
M e M(d,IR)
and
p
.
be a probability measure on the set
whose entries satisfy, for
1
< i,j,k,i < d
c Mi1i < Mk,R < c Mi.j Using Exercise 2.11 and the same route as above show that if 2
J
Log IMIIdu(M)
is finite, then the distribution of
converges to some product measure
{Log IISn x11 - n y} 2
N(O,a ) 0 V
on ]R x P(TRd)
for each nonzero x with nonnegative
coordinates. (This is rather long but requires only the former arguments, working on ILa(Bc)
instead of
L(ot)).
V.6. Large deviations
Consider a probability measure
p
on
Gl(d,IR)
such that
strongly irreducible and contracting and such that for some 1 exp T R(M) dp(M)
Tu
is
T > 0
is finite. By Theorem 4.3 we know that if we
consider the operators
T(z)
as acting on IL(a)
,
then for
Izi
small
enough
T(z) = X(z)N(z) + Q(z)
.
In particular for all real small enough
t
and unit vectors
x
130
tLog IIS xll n
IE{e
_
_ T(t)n 11(x) = A(t)°1(t) ]L(x) + Qm(t) 11(x)
}=
Since the left hand side is positive and since faster than
la(t)In
goes to zero
llQn(t)ll
is positive and
a(t)
,
tLog lISnxll
1
lim
LoglE{e
} = Log a(t)
n-*W
But
is analytic
a(t)
a'(0) = y
,
and
with
X"(0) = a2 + Y2
a
2
> 0
(see Lemmas 5.2 and 5.3). This implies the following result due to Le Page [49]
]P (Log
(a similar result holds of course also for
,
llSnxll-ny 1}
Then there exist
satisfies the
such that, for
A,B > 0
0 < e < B
lim n LogIP(Log IISnxll-ny>ne) = (c) n-_ where
q(E)
{t e - Log a(t) + ty} < 0
Sup 0 nE)
Sup O nE) >_ IP{In - zI < a}
-tZ tZ 2 IE(e n) Pn t (e n 1 IZn IIz I nE) a n Log]E(e ntz-na+n LoglPn,tZn nzI E then c(tz) - tc'(tz)->c(tE)-tEC'(tE)=inf c(t)-tE, O 0
and
lim n Log 1P(ILogIISnx n,
is finite. Then there exists x A 0
-nyI >nE) < - a
and
lim n Log ]P(ILogIISnII - n yI > nE) < - a n,_
132
Proof
:
First, replacing each
loss of generality that Let
Yn
y = 0
be the index of
p
-Y Yn , we may assume without
by
e
If
p = d
.
T
.
then the theorem follows
from the usual estimate on large deviations for sums of i.i.d. real valued random variables (see Proposition 111.1.7 and Remark 5.6). Henceforth we shall assume that in the proof of Theorem 5.4
p < d
Recall some facts we have seen
.
There exists a direct-sum decomposition
.
Ap]Rd=W®V M
such that for any An
(resp. V)
a) If
in
Gu
,
(Ap M) (W) = W
,
and, if
(Ap M) (V) = V
is a matrix which represents the action of
(resp. Bn)
ApSn on W
then the following holds.
is the distribution of
p
Al
,
then
is a strongly
T P
irreducible and contracting semigroup of Gl(r,]R)
Moreover
, where
r = dim W.
lim n Log IIAnII = Y = 0
n,
b) The upper Lyapunov exponent associated with Theorem 6.1
If we apply
and each w
W , w # 0
in
to
(A
n
)
(B n)
is negative.
we know that for some
(x> 0
< -a
(7)
,
UE -1 n Log]P(ILog IIAnwilI >n E:)
.
n
If we write this relation for a basis of
W we readily see that
1
Tim n Log ]P(Log IIAnil > n e) < -a .
(8)
n--
Since the upper Lyapunov exponent associated with m
B
n
is negative, for
large enough ]E(Log IIBmII) < 0 . Making use of the relation
Log IIBnII n e) < - a
.
(9)
n4co
Since we may choose
IIApSnII = Sup(IIAnII IIBnII)
,
(8) and (9 ) yield
Tim n Log]P(Log IIApS.II>n e) < - a and since (see e.g. Lemma 111.1.4) the index of
(Log I I Ap Sn 11 - p Log 11 Sn 11)
Log]P(Log IIS.II > n e) < -a
lim n
T11
being
p
is uniformly bounded we have (10)
133
This implies that for each
x # 0
in IRd
-a
limn Log]P(Log IISnxII > ne)
_
IIAnwII
LogIP(Log1IAPSn (xAx2A...Ax
P
,
)11y . But the spectral radius of each matrix in
S
is one. Therefore this
relation leads to
<M1 M M2 y , Since {u 1
T
=
Y>
M E T)
1
is irreducible we can find
= MI y , a .. , ud = Md y } and
basis of
t
]R
.
{v1
(15)
.
M1, M2,..
, M2d
... ,
= Md+1 y '
in
By (15) we obtain that the set {<M u. , v,>
is finite, so that
is closed and so invertible and
S = T
T
;
1
In particular the rank-one projection
.
are two
1}
denotes the
1il
x
is tight on
i=1
]Rd. But it is easily seen (look at the proof of Lemma 1.3.5) that each
limit point of this sequence is a IRd. The proposition yields that
}i-invariant probability measure on n converges to 60. Therefore E U x n 1 i=1
for any
c > 0
,
n
lim n4-
1P(Log IIS, xll < - c) = 1
E
i=1
1
which entails that the sequence
Log llSn xll
is not tight.
In particular, applying the above corollary to the sequence {e-Y Yn, n >_ 1}
we obtain the following which proves Lemma 5.3
Suppose that
COROLLARY 8.6.
Gli
is irreducible and that, for some
the sequence 1E{(Log Ids xII- n Y)2} , n = 1,2,... is bounded. n Then there exists a compact subgroup K of G1(d,]R) such that e Y Y1 x ¢ 0
,
is a.s. in K
V.9.
.
Complements
:
linear stochastic differential equations
We now consider briefly the continuous time analogue of the random products
Sn = Yn ... Y1
.
Namely
141
A random process
DEFINITION 9.1.
with values in
{St, t ? o}
Gl(d,1R)
is called a Process with Stationary Multiplicative Independent Increments (P.S.M.I.I) if the following hold is continuous in probability.
(i) (St)
(ii) For any S-11
St
... ,
tI < t2 < ... < to , the random matrices
0 Stl
St
are independent.
St
1
t
2
n n-1 Stl (iii) The distribution of St+s 1
In particular, if t ? 0, s ? 0, ut+s and by
t ? 0}
is the distribution of
ut
We denote by
- Pt * us (resp. G
T11
depends only on
s >_ 0
St , then for any
the collection
p
,
{ut
the smallest closed subsemigroup (resp.
for all G1(d,IR) which contains the support of ut It is straightforward to check that all the previous results,
subgroup) of
t ? 0.
which were proved for the discrete time model, are in fact valid for P.S.M.I.I. with these definitions of
and
T
G
u
this
moment, then for any
x # 0
{Log JISt xll, t > 0}
,
For instance if
.
P St
is strongly irreducible, and if some
Gu
has an exponential satisfies a central
limit theorem, and a large deviation estimate (one may use Exercise 9.4 in order to prove this).
An important application (see e.g. Arnold and Kliemann [1])
is
given by the so-called Linear Stochastic Differential Equations on IRd Namely let
Ao, All
{(Bt ,...,Bt)
,
t
..
,
be arbitrary matrices of order
Ar
be the usual
__>_ O}
d
Consider the Stratonovitch stochastic differential equation on
r
dxt = Ao xt dt +
and
r-dimensional Brownian motion.
E
d
IIt
Aixt o d Bt
(17)
i=1 or, equivalently, the Ito equation
r
dxt = A'' xt dt + E A. xt d Bt
A'o = Ao
i=1
Let
from
x
{xt(x)
is a
d x d
x
t ? 0}
x0(x) = x)
(i.e.
function of
,
.
r +
E
2
A.
i=1
denote the solution of (17) which starts .
It is clear that
xt(x)
is a linear
xt(x) = Stx
In other words we can write
matrix which satisfies the following S.D.E. on
r dSt = AoSt dt + E i=1
A
i S t odBt
,
So = Id
.
where
St
M(d,IR)
(18)
142
is the stochastic flow of diffeomorphisms
{St, t > 0}
In fact
associated with (17). As such Gl(d,]R)
(see e.g. Kunita [44] or Ikeda and Watanabe [37]). Let
the Lie subgroup of by
Ao, All
..
the manifold
}t = {Pt,
Ar
,
G
.
t ? 0) ,
.
The S.D.E. (18) can be considered as a S.D.E. on
contains an open set in
This implies that
.
Since
G
{Ao, All
is the closure
Gp
..
,
G . Moreover if
Ar}
,
then it is
G
(see Lemma 7.1
G
of
is connected it is easy to see that
irreducible if and only if that
takes its values in
St
ut is the distribution of St
where T11
be
G
whose Lie algebra is the one generated
Gl(d,iR)
Therefore
easy to see that
[8])
is a P.S.M.I.I. with values in
(S t)
of
G G11
is
strongly
is irreducible, which amounts to say
G
acts irreducibly on ]Rd
Besides, note that
St)-1/d
Mt = (det
satisfies the equation
St
r dMt = Ao Mt dt +
Ai Mt o d Bt
E
1=1
Ai = Ai - d (trace Ai) Id
where
,
and that r
d(det Mt) _ (trace Ao) det Mt dt + E (trace A.) det Mt o d Bt i=1 r
det Mt= Exp{(trace Ao)t + E
so that
(trace Ai) Bt} d
i=1
and
This gives that ]E(Log det Mt) = ttrace Ao
yi = trace Ao
E
i=1
Finally using well knownestimates on solutionsof S.D.E. (see e.g.
of [ ) one obtains that
Theorem IV.2.4 for each t
]E(exp2k(St))
is finite
.
Getting these results together we may apply the theorems 111.6.3, V.5.4 and V.6.2
to the process
Let
PROPOSITION 9.2.
(S t)
. We obtain
be the solution of
xt(x)
dxt = Ao xt dt +
r
A. xt
E
o d Bt
i=1 starting from
at time
x
0 . Suppose that the following hold.
(i) There is no proper linear subspace
A
i
(V)
is contained in V for all
(ii) There is no matrix Q(Ai -
1
(trace A ) Id)Q i
Q
-1 ,
i = 0,
of IRd
V 1,
..
,
r
such that
.
such that all the matrices
i = 0,...,r
are skew-symmetric.
143
Then
x # 0
(a) For all
lim
,
Log llxt(x)ll
= Y
t and y > d trace A. 02
(b) For some
{7 (Log ll xt (x) II - t y) gaussian measure
,
and all
> 0
x # 0
t > 0} converges in distribution to the
N(0, a2)
.
a > 0 and all x # 0
(c) For some
,
,
c>0
,
lim t LogiP(lLog 11xt(x)11 - tyl > tc) < -a
t--
Note that when (ii) is not satisfied then all the matrices
(det St) 1/d St lie in a compact subgroup of G1(d,IIt) Therefore we may find a norm on IRd such that Ix(x) 11 = (det St 1/d llxll for any lRd and t >_ 0 X In this case (a) holds but and only if
trace Ai = 0
(b) holds but
trace A ,
y = ,
d i = 1,...,r
,
a = 0
if
and (c) always holds.
We have of course much better results when
T11
is contracting.
To check this condition on examples it is worth noting that we have a concrete description of subsemigroup of t > 0}
and
T11
{Exp t Si
.
Namely,
TV
is the smallest closed
which contains all the matrices
Gl(d,]R)
t e ]R, i = 1....,r}
,
.
{Expt So,
This follows either
from Hormander's hypoellipticity theorem (see Siebert [65] and especially Mc Crudden and Wood [53]
,
Proposition 3) or from the
support theorem of Stroock and Varadhan (see e.g. Arnold and Kliemann
Suppose that
Example 9.3.
that trace
(A1.) = 0
generated by Then the
d
,
Ao, All
Ao, All --I Ar
0 < i Y2 > ... > Yd) ut-boundary for each t > 0 (see IV.4) and we have a complete are distinct
,
description of the behaviour of the solutions of (17) at infinity.
This follows readily from the fact that in this case from Proposition IV.2.3 since
Exercise 9.4.
GZ(d,1R)
Let
such that
Tu
{St, t > 0}
]E (11Sl 11 T)
and
Gu = SR.(d,1R)
contains an open subset of
be a P.S.M.I.I. with values in
is finite for some T > 0
.
G11
.
144
a. Show that there exists
c > 0
such that for any
0 < t
- 3/4 b. Deduce that for each
G1(d,]R)
such that
IIA
1
I
I
Since for any nonzero is close to
on
a > 0 11-11
S
v
x
dv(x) < W .
(1)
]Rd the distribution of the direction
in
V , this relation indicates that in somme sense
IISnX II is not too big. More precisely we shall show that for any <Snx,y> r > 0 1
lim - Log n-' nr
IISnxII
= 0
a.s.
.
I<Snx,y>I
This will imply that the previous limit theorems about actually valid for the coefficients of
IISnx11
are
Sn.
All the main results of this chapter are due to Guivarc'h and Raugi.
To get an idea of the way we shall prove (1) we first consider a degenerate situation. The i.i.d. matrices subset of
Gl(2,IR) and of the form
Yi b, Cal
Oi c
145
Y1,Y2,...
are in a compact
146
In that case
is of course not irreducible. We suppose that Y1 > Y2
Tu
y2 = ]E(LogIc1I).
Y1 = IE(LogIa1l) and
where
it is readily seen that
Tn = Yn...Y1
Setting
Tn e1
+ un e 2
e1
where c
ct
un = at1b1 +
lim sup (Ia
c ...c
lb
1
n a ..,a
n
1
-Y n-11)1/n < e Y 2_Y
0 < a
_ <s(M)el,el> <s(k(M)M')el,el>
this yields
immediately
x(MM') = x(M) + t(M) x(k(M)M')
.
We thus get the following formula, analogous to (2) x(Mn) = x(Y1...Yn) = x(Y1)+t(M1)x(k(MI)Y2)t ... + t(Mn-1)x(k(Mn-1)Yn).
To study this series we shall use the relations
IIx(M) II
< exp(2k(M))
(4)
IIs(M)xii
IIA2 M*II ; x orthogonal to e }
sup {
and
IIxII Since
x(M)
positive integers
is always orthogonal to
m
(5)
IIM* ells
1
and
el
these imply that for all
n ,
IIt(Mn+m)e1-t(Mn)e1II = IIx(Mn+m )e1
-
x(Mn)e1!I
n+m-1 E
IIt(Mj.) x(k(Mj.)Yj+1)II
J=n II A2 M* II2i(Y j
E ]-n JIM* ellI2
Now, the exponents associated with same, we know that a.s.
e
(Yn)
+d +1
and
(6)
(Y*)
being the
1 52
I A2 M. J *
lim 1 Log
j-
JIM* e1
II
2 = lim1 Log IIA2Y*...YIII - 2 lim ]
11 2
J
1 Log 11 Y* ...Yl el II J J
_ (Yl + Y2) - 2 Y1
=Y2 -Yl . Therefore for each
we can find
e > 0
A2 Mi
=
is finite
]P(Log R(Yj+1) > j e)
Ni
(Y2-Y1+e)
0
such
j> N2
exp 2R(Yj+l) < exp 2j e . Hence we obtain for
and any m > 0
n > Sup(N1,N2)
j(y2-yl+3c)
IIt(Mn+m)el - t(Mn)el II z
and, since
is the limit of
Proposition 1.3)1
e
E
J=n
t(Mm+n)el
as
m
(see
n(Y2-Y1+3e)
IIt(Mn )e 1 -
Z
II
e
=
(Y2-Y1+3e)
1-e This yields that for each
e > 0
a.s.
,
s(M )e
n Log II<s(Mn)el,el>
n
II
Y2-Yl+3e
The theorem will follow once we have proved the relations
(4) and (5).
On the one hand we have (see the beginning of the proof of Proposition 1.3)
IIs(M)e 111 IIx(M) II
{IIA2 M II
j=o
lE{e2a 2(Y j+1) 2
IIM3 e
II
}
}a
2a 2(Y ) {e
=
Therefore a
1
-I<x,y>Ia
_ 2a 2(Y dv(x) = 3E{--Z-i_)a
1
r Log
n
a.s.
= 0
ISnxli
Proof : We shall show that for
large enough
S
IISnxli
1
= 0
lim
n-co
Y11 Y2
satisfies the hypotheses of the above
u
r > 0
lim
n-.
where
Sn = Yn...Y1
matrices whose distribution
a.s.
,
,
(8)
<Snx,y>
n
which implies the proposition. By the Borel-Cantelli lemma it suffices to prove that for any
c > 0
S x
n
IP(I
I < c
n_S)
IR
:
by
if 0< t < E n
1
if E < t < 2e n -a if 2c n-S< t IR
n (x) = fn (I
1) IIx11
we have for 11x11 = IIx' II = 1 (I<x,y>I- 1<x',y>I)ns
,n (X-) - n(x') I
E
S
2n
a(X,X')
(see the beginning of the proof of Lemma V.4.2). This entails that, with the notations of Chapter V, 110,111a =
ma(cn) 6
2e
S
+1
1 58
for any
a
a
for
0 < a < 1
,
Sn E
Taking into account Theorem V.2.5
.
we obtain,
small enough,
x
y>I < en-S) < E
1P(I
I
e,>I
LogI(Sn)1,jI - LogI(S.)i'jl = Log
Log
n s*
IIf*le JII
n
II
's*
e
n
= Log i<e
>I-LogI<e,n
,
1
i
II snej II
J
>I
I I Snej II
converges in probability to an a.s. finite limit (this follows from the fact, seen in Theorem 111.4.3, that a random direction
Z
satisfying
converges in probability to
n
j
IP( = 0) = 0
for any nonzero x).
Therefore (Lo g
n Exercise 2.4.
I
n 1,1I- Lo g
(S )
i
Suppose that
contracting and that for some Show that for any nonzero to
-yd
(S
n
Tu'
)
i1jl)
converges to 0 in
is strongly irreducible and
T > 0
,
J exp T i(M) d}i(M)
x , 1 Log IIS-lx II
n
probability.
n
is finite.
converges almost surely
.
VI.3. Behaviour of the rows
As the preceding results, the following is due to Guivarc'h and
1 60
Raugi (see [59]). We use the notation
Suppose that
THEOREM 3.1.
and that x
is finite for some
converges to
,
.
is strongly irreducible and contracting
Tu
exp T Z(M) d}i(M)
f P(]Rd)
in
1.2
T > 0
Then for any
.
Using the
a.s.
Z
converges to
Proof : We first show that
.
Z
notations of the first section we have d
= M el
Y ...Y e
n
1
n
1
s(M ) k(M )e
n
1
n
= 1
s(M )e. n 1 1' 1
Hence
s(n)e1 <s(M )e1'e n
M1 e1 <M
n e 1' e I >
s(n)ei
d
+
> 1
E
n e1'e1>
<M
i =2
Taking into account (5) and (i) of Corollary 2.3 we see that, for
i = 2, lim
,d l n
,
I s (Mn)eill Log
n-*
I<Mnei,el>I
lim
Sn = Yn...Y1
then
Y11Y2....
we find
converges a.s. to some direction
(this improves Theorem 111.4.3). Choosing
x
equal to some basis
Z
1 61
we see that the directions of the rows of e. 1 a.s. to the same limit. vector
Sn
converge
VI.4. Regularity of the invariant measure
We shall now see, following Guivarc'h and Raugi, that although the p-invariant distribution
is often singular, it always has some
v
regularity.
Suppose that
PROPOSITION 4.1.
T11
Then there exists
on
is contracting, strongly
T > 0
irreducible and that, for some
is finite.
f exp T R(M) dp(M)
,
p-invariant distribution
such that the
a > 0
v
P(]Rd) satisfies -
Sup d
and
u
dv(x)
2 2 <x,u>2
.
Hence 11-11
Il
I<x,u>l
llxAy11 Since, by Theorem 1.2, there exists some
Sup
f(
lxII {
the proposition is clear
-
}a dv(x)
0
such that
I<x,u>
Ilu 11=1
COROLLARY 4.2.
a > 0
x
in
P(]Rd) and
v({Y ; 6(x,Y) < E}) < C
Ea
x
and radius
v(B) = J
a(x,Y)
B 6(x,y)a
e d(x,Y)-a
dv(Y) < E:(' 1
and
.
Proof : This follows readily from the fact that if center
a > 0
e > 0
dv(Y)
B
is the ball with
1 62
As a consequence let us prove that the dimension of
V
is
positive. First recall the definition of the Hausdorff dimension of a
A
subset
(P(1Rd),6). The reader is referred to
of the metric space
Billingsley [5]
for details. For any
I > 0
and
E > 0
let
ks(A,E) = inf E (diam B.) i
where the infinimum extends over all countable coverings of A by closed balls
of diameter less than
B,
i
E
,
and let
R(A) = lim R,(A,E) E4O
A
The Hausdorff dimension of
dim A = sup{(
is defined by
i (A) = W} = inf{ : R(A) = 01
:
13
dim V = inf{dim A ; A is a Borel subset such that V(A) J01. Now if
v(A) # 0
E v(Bi)
V(A) Za(A) >_ v(A)
so that
and if
and
(B
0.
According to Ledrappier [46] one may conjecture that, for
2 x 2
matrices,
dim v = - 2y f
y) (x) dv(x) du(g)
Log d(B
(9)
(This is actually proved in [46] but for a weaker notion of dimension).
We have seen (see 11.7.2) that plane
H = {z E
IC
;
Imz > 0}
g.z when
g = [ a
d
]
.
is just the action of 1R U {-} p
, then
If
.
V
GP
Sl(2,1R) acts on the upper half
by the formula
az + b cz + d
Also that the extended action on the real axis Sl(2,]R) on
P(1R2)
if we identify
P(1R2) with
is the closed subgroup generated by the support of
is carried by the so-called limit set
L = {X E ]RU {oo} ; ](gn) C G11
L
of
Gu
,
i.e.
, his g n i = X} n_ a_
(This follows from the fact that
L
is closed and
G
- invariant or
u
1 63
converges
directly from Corollary 11.7.2 which asserts that to a random variable with law
V)
.
L may be arbitrarily small
It is known that the dimension of
V may be (and is often)
(see e.g. Beardon F3]). In particular singular.
It is worth rephrasing the above corollary in the context of projective transformations.
Let
COROLLARY 4.3.
G1(2,1R) such that
be a probability on
1i
strongly irreducible and contracting. If for some eT £(M)
Tu
is
,
is the probability measure on
is finite and if v
dp(M)
T > 0
]R
f
which satisfies
J
f(ax+b) dp(a,b,c,d) dV(x) = J f(x) dv(x) cx +d
for any bounded continuous function
a > 0
such that
C > 0
and
]R, then there exist
on
f
(i) For any
in
y
]R
J Ix-yl-a dv(x) < C (
(ii)
J
(1+x2)a/2 dv(x) < C
.
In particular the distribution function
F(x) = v(--,x]
is H6Zder
continuous.
Proof and
:
The mapping
'Y(°°)
= e1
]R U {oo} > P(]R2) defined by P(x) =xe1 +e2
T :
sends
v
onto the unique
P(]R2). Therefore, by Proposition 4.1 such that for any real
,
p-invariant measure on
there exist
y d(T(y),T(x))-a
dv(x)
R
and
( J
]R
But
d(1(y),Y(x))
II (x e1+e2) A (yet +e2) II Ix e1 +e 211
Ily et + e 211
C > 0
and
a > 0
1 64
Ix-y I /1+J
and
6(1'('),T(x)) =
1+yZ
I(Xe1 +e2) A e1 II
1
1+xZ
1+xZ The proposition follows immediatly.
satisfies a law
v
Finally we notice that the probability measure
of pure type with respect to the natural measure md . Recall that
]R U [-I, m2
On the projective line
and
d
P(]R ),(see II.5.1).
is the unique rotation-invariant distribution on
can be identified with the Cauchy
distribution and is equivalent to the Lebesgue measure.
PROPOSITION 4.4.
Let
be a probability measure on
p
has only one p-invariant distribution v
- or
V
and md
- or
v
is purely discrete.
Proof : Write
component of
v
on
P(IR)d). Then
is absolutely continuous w.r.t.
-either
for the discrete (resp. continuous) P * V2
It is readily seen that
.
and
are mutually singular ,
(resp. v2)
V1
V
G1(d,]R) which
is continuous.
Therefore the equality *
V = V entails that thus
.
But
v1 + u
V2
and
11 * v2
have the same mass,
By uniqueness either
V = V2
or
v2 > }1 * v2
V2 = 11 * V2
*
V = u
Suppose now that
V
.
v2
is continuous and write
V = V1
Va.c
.
(resp. V ) for s
its absolutely continuous (resp. singular) component w.r.t.
for any M
Gl(d,]R)
in
Mmd is equivalent to
and
m
.
d
Since
(see 11.5.1)
(
u * Va.c =
I
M Va.c dp(M)
is absolutely continuous, entailing as above that
11 * Va. c
uniqueness of the continuous or
Example 4.5.
V
p-invariant measure either
V
= V
a.c
By
is absolutely
is singular.
It was already noticed by Kaijser in [38] that is it easy
to exhibit singular invariant measures. Consider for instance two non negative
2 x 2
matrices
1 65
[ai
bt
c,
dt
M1
Idet Mil = 1
such that
show that if
a2
1 d2 J
L c2
M2
ci > 0
,
b2
and
> 2
d
i = 1,2
for
.
Let us
is the Bernoulli measure defined by
P
u({M1 }) = P({M2}) = 1/2 then the
11-invariant distribution
We shall consider
V
on
prove that the Lebesgue measure of the support For
i = 1,2
is singular.
P(1R2)
]R U {-} and
as a measure on the projective line
V
and
x
in IF, U {m}
V
of
S
is zero.
set
a.x + b. i
q.(x) =
i
c.x + d.
i
a2
a = Max(a1,
It is readily seen that if
.([O,a]) under
is contained in
T.
,
[O,a]
implying that
Therefore
.
(x) I =
.
Now, since for
0
2 , m(A) = 0
m(A)
Exercise 4.6.
Let
11
be a probability measure on
a unique invariant distribution v
V
is the smallest closed subset
M
Exercise 4.7. (or
[O,a]
x ? 0
any
of
then
[0,a]- is invariant
is contained in
S
bt
b2) c2' d1' d2
c1
in
Tu
Gl(d,]R) which has
P(]Rd). Show that the support
on S
of
P(]Rd) which satisfies
.
(a) Suppose that
P
is a probability measure on
G1(d,]R)
S1(d,]R)) which has a bounded density with compact support w.r.t.
166
the Haar measure. Show that (b) Deduce that if
has a bounded density.
V
is absolutely
V
is not singular then
P
continuous.
Show that the Cantor distribution is
Exercise 4.8.
some non-irreducible probability measure
VI.5. An example
Let
(5.1)
:
p
p-invariant for
Gl(2,IR).
on
Random continued fractions
{An, n
1}
>_
be a sequence of independent random
positive integers with a common distribution
p , defined on some
1P). Consider the 2 x 2 matrices
(0, A
0
[1
Yn and call
p
A
n
their distribution. For any
in the projective line
x
]R U {co} , 11 Al + A2+ 1
+An+x Notice that Z
with values in
converges a.s. to a real random variable [0,1]
The distribution
.
V
of
is the
Z
p-invariant distribution on the projective line. This follows readily from the fact that the
A n's
are positive integers and from the
elementary properties of continued fractions (see e.g. Billingsley [5]). (Much less obvious is the fact that this also holds for arbitrary real i.i.d.
A
n
under an integrability assumption, which is an immediate
consequence of Theorem 3.1). Let us show that
V
is singular as soon as the
a.s. constant (in this case
Tu
A,'s i
are not
is strongly irreducible and
contracting, cf. Exercise 5.4). This result is due to Chatterji C1?. First recall that any real number
t
in
(0,1)
may be expanded
in a continued fraction, i.e. t = lim (a1 (t)
n-
where we define inductively
;
a2(t)
...
;
an(t))
1 67
1
(a 1) =
When
a1+(a2;...;an)
is irrational this expansion is unique. In particular since
t
A1(w),...,An(w)
law of
V
V-a.s. unique. As a consequence the
is continuous, the expansion is
under
1
(at;...;an) =
a1
under
is the law of
]P
a1(t),...,an(t)
V
Consider now the mapping
T
(0,1) -> (0,1)
:
defined by
T(t) = t - [t] Since T
(a1(T(t))
;
a2(T(t))
;
a3(T(t))
;
...) = (a2(t)
;
a3(t)
;
...)
,
is equivalent to the shift on the sequences of integers. The 0-1
law thus implies that
is invariant and ergodic under
V
(1+x)-1
well known that the measure under
T
dx
on
(0,1)
(see e.g. Billingsley). Therefore either
w.r.t. the Lebesgue measure or a1(t), a2(t), ...
v = 1
(0,1)
(Log 2) -1
.
But it is
is also ergodic V
is singular
(1+x)-1 dx
.
But
are not independent under this latter distribution
(elementary computations). This proves that
(5.2)
T
V
is singular.
Let us outline a proof of Ledrappier's formula (9) in this
simple situation, which is in this case due to Kinney and Pitcher [43]. Let
Jn(w) = Y1 (w)...Yn(w) 0,1] It is readily seen that we may write
Mn = Y1...Yn =
where
qn = Angn-1 + qn-2
[Pn_i
Pn
L qn_1
qn
The length .
IJn(w)I
of
Jn(w)
IJn(w)I = IMn(w)'0 - Mn(w)'1I =
But, since
q
n
+
qn2 Therefore, if
'
2 qn-1
]E(Log A1)
associated with
n,. n
{gn_I(w)(gn(w)+qn-1(w))}-
5 q n-1
2
lim
satisfies
(Yn)
15
gn-1(gn
+q n-1)
2 + 2 - qn-1 qn
is finite and if
y
is the Lyapunov exponent
then
Log IJn(w)I = - lim n Log IIMn(w)e2112 = -2y
On the other hand
IF a.s.
168
V(Jn(W)) = V({t e (0,1)
n II
p({A()})
i=1
i
(Recall that under
V
a1(t),...,an(t)
,
distributed according to if
p(i) = pi
a1 (t) = A1(u),...an(t) = An(W)})
;
are independent r.v.
p). Making use of the law of large numbers,
,
n
n Log V(Jn(W)) = 7E (Log p(A1)) E
pi Log pi
a.s.
i=1 (Note that since
Log x < x-1
E pi Log i = ]E(Log A1)
and since
is
finite
IEpiLogpil =EpiLog Z + 2EpiLogi < (E 2)-1 + 2 1E (Log A1) i pi i is also finite).
Therefore
Ep Logp
Log V(Jn (W) )
i
lim
n- Log J. (W) = In other words if for any
in
t
In(t) = is a (0,1)
;
2y
i
]P
a. s.
we set
(0,1)
a1 (s) = a1 (t) ,...,an(t) = an(s) }
then
Log V(In (t) )
him n-
Ep. Logp. 2Y
Log IIn(t)
1
V-a. s.
Onecan deduce from this formula (see Kenney and Pitcher [43]) that
dimV= -
E pi Log pi 2y
In order to recover Ledrappier's formula (9) it suffices to note that if
M = (0
1)
then n
\)(M-In(t)) = p(a)
p(ai(t)) = p(a) V(In(t))
ll
i=1
(use the formula
V({s ; ai(s) = a, a2(s) = al(t),
an+1 (s) = an(t)}))
Thus
d (M
N))
= p (a)
J Log d(M dN)
on
(0,I)
and
1V) dV dp(M) = E pi Log pi
169
An interesting example of such a measure is the Minkowski
(5.3)
measure denoted by distribution on
(0,1). We may define it as the
on
?
]R U {oo} where
(0
n)
=
11-invariant
is defined by
p
n
I
n
,
2
In order to describe it let us notice that if
is the Bernoulli
p
measure which verifies i(A) = 11(B) = 1 /2 for
B=( then
0)
p-invariant. (This is proved as follows
is also
?
B=(
11)
ti
:
if
V
is
ti
p-invariant then
Z(6A+6B) * V =v
v=26A *v+2 6B * (26A+Z dB)
and
V
2 6A * v+ 4 6BA * v+ 4 6B 2* v By induction we have for any
n > 0
,
n
V= E - 6 m=1
and, going to the limit
v But
m=1
*v
d
Bn
2n
,
m=1 2m
E L 6
p =
* V+
Bm-1A
2m
V
dBm-1A
, hence
V = ?
by uniqueness of the invariant
Bm-1A
2m
measure).
We now define inductively the Brocot sequence of order p1(n) p2 ( )
q1 (n)< - .. - 0
]E{IAIT} is finite. Show that this is also true when A integer-valued, as soon as
Exercise 5.5.
[11]
distributions on
[51]. For
]E(LogIAI)
a > 0
,
11
is positive-
is finite.
let
and
as
]R+ defined by
aa(dx) = 2 exp - 2 1 0
dx
va
is
such that
be the two
171
2 (x + X) 1(0,co) dx
va(dx) =
where
j ur-1 exp -
Kr (a) =
a
(u + u) du
0 2
a. Consider two independent random variables the law of
X
(resp. A)
is
va
(resp. Aa)
.
X
and
A
such that
Show that the law of
A)- 1
(X +
is
Va
.
(Hint
:
compute the Laplace transforms).
b. Show that the Lyapunov exponent distribution of
0 (1
1
A)
y
2Ko(a) is equal to
aK
1 (a)
associated with the
SUGGESTIONS FOR FURTHER READINGS
We provide some recent references on limit theorems for random matrices and related topics. We make no claim for completeness but the quoted papers often contain a large bibliography.
(1) Lyapunov exponent for stationary sequences
The main properties of Lyapunov exponents in the stationary case can be found in Ledrappier [46] . Guivarc'h [32] (see also Royer [62] Virtser [69],[70]) gives a criterion ensuring that two given exponents are distinct for markovian products. See Ledrappier [48] for an assumption implying that the exponents are not all equal, in the general stationary setting.
(2) Boundary theory
After the fundamental work of Furstenberg (see [20] , the set of bounded harmonic functions was determined
E221, [23] )
- for absolutely continuous distributions on connected groups by
Raugi [59]
(see also Guivarc'h E311),
- for distributions on discrete groups of matrices by Ledrappier
[47]
.
(3) Limit theorems (3.1)
Onecan find a proof, under our usual irreducibility
assumptions, of - the functional central limit theorem, - the law of iterated logarithm, - the renewal theorem,
173
174
- the local limit theorem,
for the sequence
Log II Sn x II
in Le Page [49], [50]
properties are studied in Guivarc'h [33] The central limit theorem for
Sn
.
Recurrence
(see also Bougerol [9]) written iri the polar and the
Iwasawa decomposition is proved in Raugi [59]
.
References to earlier
proofs and applications can be found in Tutubalin [68]. (3.2)
Properties of the solutions of the difference equation on
d IR
Xn+1 = Yn Xn + Bn Bn
(where
[41] d = 1
is in
IRd
and
and in Le Page [50] ,
Grincevicius [34]
(3.3)
.
Yn
in
G1(d,1R))
are studied in Kesten
Stationary solutions are given in [8]
.
For
proves a central limit theorem.
Without irreducibility assumptions, the central limit
theorem is not yet fully understood. The latest reference is Raugi [60].
(4) Positive matrices
The reader will find in Cohen [14] a nice account of the applications of products of positive random matrices to demography and an extensive bibliography. Kesten and Spitzer [42] study the convergence in distribution of such products.
(5) Linear stochastic differential equation
A goog introduction to this subject is the survey of Arnold and Kliemann [1]
. A nice application is given in Pardoux and Pignol [58].
BIBLIOGRAPHY
[1] ARNOLD, L. and KLIEMANN, W. (1983). Qualitative theory of
stochastic systems. In "Probabilistic Analysis and Related topics", A.T. Bharucha-Reid (ed.), Vol. 3, 1-79, Academic Press, New York.
[2] ARNOLD, L., CRAUEL, H. and WIHSTUTZ, V. (1983). Stabilization of linear systems by noise. SIAM J. Control Optim. (21), 451-461. [3] BEARDON A.F. (1966). The Hausdorff dimension of singular sets of properly discontinuous subgroups. Amer. Journ. of Math. (88), 722-736.
[4] BELLMAN, R. (1954). Limit theorem for non-commutative operations. I. Duke Math. J. (21), 491-500. [5] BILLINGSLEY, P. (1965). Ergodic theory and information. Wiley and Sons. New York.
[6] BIRKHOFF, G. (1957). Extensions of Jentzsch's theorem. Trans. Amer. Math. Soc. (85), 219-227.
[7] BOUGEROL, P. (1984). Stabilite en probabilite des equations differentielles stochastiques lineaires et convergence de products de matrices aleatoires. C.R. Acad. Sc. Paris, (299), Serie 1, 631-634.
[8] BOUGEROL, P. (1984). Tightness of products of Random matrices and stability of linear stochastic systems. To appear in Ann. Probab.
[9] BOUGEROL, P. (1985). Oscillation des produits de matrices aleatoires dont 1'exposant de Lyapounov est nul. To appear. [10] BREIMAN, L. (1968). Probability. Addison Wesley. Reading Massachusetts. 175
176
[11] CHASSAING, P., LETAC, G. and MORA, M. (1984). Brocot sequences and random walks on
Sl(2,1R). In "Probability measures on
groups 7", ed. H. Heyer. Lecture Notes in Math. n° 1064. Springer Verlag. Berlin, Heidelberg, New York, 36-48.
[12] CHATTERJI, S.D. (1966). Masse, die von regelmassigen Kettenbruchen induziert sind. Math. Annalen (164), 113-117. [13] CHEVALLEY, C. (1951). Theorie des groupes de Lie, t.2
:
groupes
algebrigues. Hermann, Paris. [14] COHEN, J.E. (1979). Ergodicity theorems in demography. Bull. Amer. Math. Soc. (3), 275-295.
[15] COHEN, J.E. and NEWMAN, C.M. (1984). The stability of large random matrices and their products. Ann. Probab. (12), 283-310. [16] DEKKING, F.M. (1982). On transience and recurrence of generalized random walks. Zeit. fur Wahrscheinlichkeitstheorie and Verw. Gebiete. (61), 459-465. [17] DUNFORD, N. and SCHWARTZ, J. (1958). Linear Operators, Vol. 1. Interscience.
[18] FELLER, W. (1971). An Introduction to Probability Theory and its Applications. Vol. 2. Wiley. New York, London, Sydney, Toronto. [19] FREIDLIN, M.I.
and VENTSEL, A.D. (1984). Random perturbations of
dynamical systems. Springer Verlag. Berlin, Heidelberg, New York.
[20] FURSTENBERG, H. (1963). A Poisson Formula for semisimple groups. Annals of Math. (77), 335-383. [21] FURSTENBERG, H. (1963). Non-commuting random products. Trans. Amer. Math. Soc. (108), 377-428.
[22] FURSTENBERG, H. (1971). Random walks and discrete subgroups of Lie groups. Adv. Proba. and related topics (1), 3-63. [23] FURSTENBERG, H. (1972). Boundary theory and stochastic processes in homogeneous spaces. In "Harmonic analysis on homogeneous spaces", Symposia on Pure and Applied Math. Willamstone, Mass. [24] FURSTENBERG, H. (1980). Random walks on Lie groups. In "Harmonic analysis and representations of semi-simple Lie groups". J.A. Wolf, M. Cahen and De Wilde (eds.). D. Reidel Publishing
1 77
Company. Dordrecht, Holland. [25] FURSTENBERG, H. and KESTEN, H. (1960). Products of random matrices. Ann. Math. Statist. (31), 457-469. [26] FURSTENBERG, H. and KIFER, Y. (1983). Random matrix products and measures on projective spaces. Israel J. Math. (10), 12-32. [27] GLASNER, S. (1976). Proximal Flows. Lecture Notes 517. Springer
Verlag. Berlin, Heidelberg, New York. [28] GORDIN, M.I. and LIFSCHITZ, B.A. (1978). The central limit theorem for stationary Markov processes Soviet. Math. Dokl. (19), N° 2, 392-394.
[29] GREENLEAF, F. (1969). Invariant means on topological groups. Van Nostrand. New York. [30] GRINCEVICIUS, A.K. (1974). A central limit theorem for the group
of linear transformations of the real axis. Soviet
Math. Dokl.
(15), 1512-1515.
[31] GUIVARC'H, Y. (1980). Quelques proprietes asymptotiques des produits de matrices aleatoires. In "Ecole d'e"te de SaintFlour 7 - 1978", ed. P.L. Hennequin.Lecture Notes in Math. n° 774. Springer Verlag. Berlin, Heidelberg, New York. [32] GUIVARC'H, Y. (1984). Exposants caracteristiques des produits de
matrices aleatoires en dependance markovienne. In "Probability measures on groups 7", ed. H. Heyer. Lecture Notes in Math.
n° 1064. Springer Verlag. Berlin, Heidelberg, New York, 161181.
[33] GUIVARC'H, Y. (1984). Application d'un theoreme limite local a la transcience et a la recurrence de marches de Markov. In "Theorie du Potentel". Lecture Notes in Math. n° 1096. Springer Verlag. Berlin, Heidelberg, New York, 301-332. [34] GUIVARC'H, Y. and RAUGI, A. (1985). Frontiere de Furstenberg, proprietes de contraction et theoremes de convergence. Zeit. fur Wahrscheinlichkeitstheorie and Verw. Gebiete. (69), 187242.
[35] HENNION, H. (1984). Loi des grands nombres et perturbations pour des produits reductibles de matrices aleatoires. Zeit. fur Wahrscheinlichkeitstheorie and Verw. Gebiete (67), 265-278.
1 78
[36] HEWITT, K. and ROSS, A.
(1963). Abstract Harmonic Analysis 1.
Springer Verlag. Berlin, Heidelberg, New York. [37] IKEDA, N. and WATANABE, S.
(1981). Stochastic differential
equations and diffusion processes. North Holland-Kodansha Amsterdam, Takio.
[38] KAIJSER, T. (1970). On distribution problems for random products of non-commutative matrices. Report n° 22. Uppsala University. [39] KAIJSER, T. (1972). Some limit theorem for Markov chains with
applications to learning models and products of random matrices. Report Institute Mittag-Leffler, Djursholm, Sweden. [40] KAIJSER, T. (1978). A limit theorem for Markov chains in compact
metric spaces with applications to products of random matrices. Duke Math. Journ. (45), 311-349. [41] KESTEN, H. (1973). Random difference equations and renewal theory for products of random matrices. Acta Math. (131), 207-248. [42] KESTEN, H. and SPITZER, F. (1984). Convergence in distribution for products of random matrices. Zeit. fur
Wahrscheinlichkeitstheorie and Verw. Gebiete. (67), 363-386. [43] KINNEY, J.R. and PITCHER, T.S. (1966). The dimension of some sets
defined in terms of f-expansions. Zeit. fur Wahrscheinlichkeitstheorie and Verw. Gebiete. (4), 293-315. [44] KUNITA, H. (1984). Stochactic differential equations and stochastic flow of diffeomorphisms. In " Ecole d'ete de
Saint-Flour 12-1982", ed. P.L. Hennequin. Lecture Notes in Math. n° 1097. Springer Verlag. Berlin, Heidelberg, New York. [45] LANG S. (1965). Algebra. Addison-Wesley. Reading, Massachusetts. [46] LEDRAPPIER, F. (1984). Quelques proprietes des exposants caracteristiques. In "Ecole d'ete de Saint-Flour 12-1982", ed.
P.L. Hennequin. Lecture Notes in Math. n° 1097. Springer Verlag. Berlin, Heidelberg, New York. [47] LEDRAPPIER, F. (1984). Poisson formula of discrete groups of matrices. To appear in Israel Jour. Math. [48] LEDRAPPIER, F. (1985). Positivity of the exponent for stationary sequences of matrices. To appear.
179
[49] LE PAGE, E. (1982). Theoremes limites pour les produits de matrices aleatoires. In "Probability measures on groups", ed. H. Heyer. Lecture Notes in Math. n° 928. Springer Verlag. Berlin, Heidelberg, New York, 258-303. [50] LE PAGE, E. (1983). Theoremes de renouvellement pour les produits de matrices aleatoires. Equations aux differences aleatoires. Seminaire de Probabilites. Universite de Rennes. [51] LETAC, G. and SESHADRI, V. (1983). A characterization of the
generalized inverse Gaussian distribution by continued fractions. Zeit. fur Wahrscheinlichkeitstheorie and Verw. Gebiete. (62), 485-489.
[52] LOOMIS, L.H. and STERNBERG, S. (1968). Advanced Calculus. Addison Wesley. Reading, Massachusetts. [53] Mc CRUDDEN, M. and WOOD, R.M. (1984). On the supports of absolutely
continuous Gauss measures on
Sl(2,1R). In "Probability
measures on groups 7", ed. H. Heyer. Lecture Notes in Math.
n° 1064. Springer Verlag. Berlin, Heidelberg, New York, 379397.
[54] MOSTOW, G.D. (1955). Self-adjoint
groups. Ann. of Math. (62), 44-
55.
[55] NAGAEV, S.V. (1957). Some limit theorems for stationary Markov chains. Theor. Proba. Appl. (2), 378-406. [56] NEWMAN, C.M. (1984). The distribution of Lyapunov exponents Exact results for random matrices. To appear. [57] OSSELEDEC, V.I. (1968). A multiplicative ergodic theorem. Trans. Moscow Math. Soc. (19), 197-231. [58] PARDOUX, E. and PIGNOL, M. (1984). Etude de la stabilite de la solution d'une E.D.S. bilineaire a coefficients periodiques.
Application au mouvement des pales d'helicoptere. In "Analysis and Optimization of Systems, Part 2"
;
A. Bensoussan and J.L.
Lions. (Eds.), Lecture Notes in Control and Info. Sci. 63. Springer Verlag.
[59] RAUGI, A. (1977). Fonctions harmoniques et theoremes limites pour les marches aleatoires sur les groupes. Bull. Soc. Math. France, Memoire 54, 127 p.
180
[60] RAUGI, A.
(1980). Quelques remarques sur le theoreme de la limite
centrale sur un groupe de Lie. C.R. Acad. Sc. Paris, (290), 103-106.
[61] REVUZ, D. (1984). Markov chains. North Holland. Amsterdam, New York, Oxford.
[62] ROYER, G. (1980). Croissance exponentielle de produits markoviens de matrices aleatoires. Ann. I.H.P. (16), 49-62.
[63] SAZONOV, V.V. and TUTUBALIN, V.N. (1966). Probability distributions on topological groups. Th. Proba. Appl. (13), 1-45.
[64] SENETA, E. (1981). Non-negative matrices and Markov chains, 2nd ed., Springer Verlag. Berlin, Heidelberg, New York.
[65] SIEBERT, E. (1982). Absolute continuity, singularity, and supports of Gauss semigroups on a Lie group. Mh. Math. (93), 239-253. [66] TUTUBALIN, V.N. (1965). On limit theorems for products of random matrices. Theor. Proba. Appl. (10), 15-27. [67] TUTUBALIN, V.N. (1969). Some theorems of the type of the strong law of large numbers. Theor. Proba. Appl. (14), 313-319. [68] TUTUBALIN, V.N. (1977). The central limit theorem for products of random matrices and some of its applications. Symposia Math. (21), 101-116.
[69] VIRTSER, A.D. (1979). On products of random matrices and operators. Theor. Proba. Appl. (24), 367-377.
[70] VIRTSER, A.D. (1984). On the simplicity of the spectrum of the Lyapunov characteristic indices of a product of random matrices. Theor. Proba. Appl. (28), 122-135.
PART
B
RANDOM SCHRODINGER OPERATORS
INTRODUCTION
Numerous equations arising from one dimensional discrete physical systems lead to the analysis of a linear second order difference operator H, acting on a complex sequence 'Y
n
, n e 7L , by
(H'1' )n = b -1 ((-AY ')n + an 'fn)
In this formula, A is the discrete Laplacian (AY')n ='n+1 +T n-1 and an bn , are two fixed sequences of real numbers with bn > 0,
Tn
'
representing the physical properties of the medium. Generally such operators are associated to "time dependent" equations and we give some typical examples
:
(i) A solution of the Schrodinger equation i
-i) t satisfies a
0(n,t) = Y'n
a
at
= H of the form
HT = a'Y where H is the classical Schrodinger
operator, that is the operator associated to b
n
= 1, Vn E 2Z, and a
n
is
the potential at site n. 2
(ii) A solution of the wave equation bn
iat satisfies a
0(n,t) = 'n
a
2 - _
of the form
at
2
HT = a 'Y where H is the "Helmotz operator" that
is the operator associated to an = 0, Vn E Z Z, and bn is the diffusion coefficient at site n.
(iii) A solution of the heat equation bn (A )n of the form at = ¢(n,t) = Tn a XtsatisfiesH'Y =a4' where H is the Helmotz operator.
Similar equations and operators also appear in quasi-one dimensional systems associated to an infinite wire of finite cross section with R sites. We have only to replace the sites n by (i,n) where i e (1,
..
,
R), (the integer k is called the width of the strip) and
the real sequences an,bn by matrices sequences. It is well known that the spectral properties of the operator 11
183
184
viewed as a self adjoint operator on an Hilbert space, govern the asymptotic behavior of the solutions of the associated time dependent equation. In quantum mechanicsthe number I$(t,n)12 (normalized in a itH
of the time f) n dependent Schrodinger equation, represents the probability of presence of
suitable way) associated to a solution (t,n) = (e
a particle at the site n at time t. Roughly speaking, when T is associated to the continuous spectrum of H then we have the "diffusion" T
behavior
lim
1
T}+ 2T -T
I$(t,n) 2dt = 0, and if T is associated to the
point spectrum of H then we have the "localization" property lim
Sup t
E
I4(t,n)12 = 0.
(See D. Ruelle [53] for more details). In
n,N
deterministic systemswith periodic structure, it is known that only diffusion behavior occurs. But it has been remarked that localization appears when this regular structure is perturbed by impurities or inhomogeneity in the medium. Thus a "metallic" wire suddenly becomes an insulator. We first give below a brief historical survey of the mathematical approaches to this subject.
(1958) that for the
It was first announced by P.W. Anderson [1]
classical multi-dimensional Schrodinger operator with an independent identically distributed family of random potentials, the spectrum has to be pure point for a "typical sample" assuming the disorder "large enough". It was later conjectured by N. Mott and W.D. Twose [46]
(1961)
that in the one dimensional case, this should be true at any disorder. The works of H. Furstenberg, H. Kesten [20]
(1960), H. Furstenberg [19]
(1963), V. Osseledec [49] (1968) provided the essential mathematical
background used in the first rigorous approaches of the subject. It was first proved by H. Matsuda, K. Ishii [44]
(1970), A. Casher
J.L. Lebowitz 110] (1971), L.A. Pastur [50]
(1973), Y. Yoshioka [64]
(1973) that there does not exist an absolutely continuous component in the spectrum of H. An essential step was achieved in 1973 when I. Ja. Goldsheid, S.A. Molcanov, and L.A. Pastur [24] gave the first proof to the conjecture of Mott and Twose (they actually dealed with the continuous case). Their original proof has been later considerably simplified and extended by R. Carmona [7] [52]
(1982),
[8]
(1983), G. Royer
(1983), J. Brossard (1983). In the "discrete" case the same result
has been obtained by H. Kunz, B. Souillard [37]
(1980), J. Lacroix [38]
(1982), F. Delyon, H. Kunz, B. Souillard [13] (1983). Moreover Goldsheid
gave a similar announcement in a strip [23](1981) and the proof can be
185
found in J. Lacroix [39]
(1983) [40]
[41]
(1984).'All these previous
proofs of localization are rather technical and at times the essential guiding principles are not easily understood. Fortunately, in the late of 1984, S. Kotani [36] clarified the situation, giving a rigorous statement to an earlier claim of R.E. Borland [5]
(1963).
Our essential goal is to provide a direct and unified treatment to the foregoing problems, in the general setting of operators H introduced at the begining of this discussion here in after called Schrodinger operators. The essential tool will be the theory of products of i.i.d. random matrices developed in the first
part of this
book. We are mainly concerned with the independent case but a lot of definitions and properties are given in the general ergodic case. The "almost periodic case" is also of great physical and theoretical interest but most of the proofs have nothing to do with random matrices.
Interested readers have to look at the survey of B. Simon [57] where they can also find an extensive bibliography. Since it seems that the theory of random matrices can hardly be used in the multidimensional case (up to now,limiting procedures in strips whose width goes to infinity have not been successful) we restrict ourselves to the one dimensional case and strips.
In chapter I the essential definitions and properties related to the spectral analysis of the deterministic operator H are given. In particular we construct a sequence of "good approximations" of the spectral measure of H and establish the existence of "slowly" increasing generalized eigenfunctions. Moreover the links between the singularity of the spectrum and the fundamental notion of hyperbolic behavior of a product of matrices are pointed out.
In chapter II we define an ergodic family of Schrodinger operators which contains as essential examples the classical Schrodinger operator and the Helmotz operator. Some weak properties of the spectrum of H considered as a subset of JR are given, before introducing the essential
concept of Lyapunov exponent. Positivity of this exponent is carefuly studied since this property is crucial in order to obtain absence of absolutely continuous spectrum. The distribution of states describing
186
the asymptotic behavior of the eigenvalues distribution for the operator restricted to "boxes" is of great physical interest and we discuss in detail its regularity properties together with the links with the Lyapunov exponent (Thouless formula). Kotani's criterion insuring localization property is then introduced in the general ergodic case but its main application to the independent case is discussed in the following chapter. Finaly we give a straightforward application of the central limit theorem on SL(2,1R)
to the asymptotic
behavior of the conductance.
Chapter III is devoted to the proof of the conjecture of Mott and Twose both in classical Schrodinger and Helmotz case. In the first model, Kotani's criterion gives immediately the solution, but in the general case the proof is more involved and requires some Laplace analysis on SL(2,IR) . As a consequence extra properties of the
distribution of states are obtained.
All these foregoing results are generalized in the chapter IV to the case of a strip. Most of the previous proofs in the one dimensional case
can be translated with some care. But some problems
are much more involved, especialy positivity of Lyapunov exponents. General results given in the first part of the book are then very useful. The proof of localization in the general case requires also much more work since Laplace analysis on symplectic groups needs some knowledge about symplectic geometry.
Numerous related topics, non-stationary processes for instance, are not tackled when they don't appear as direct applications of products of i.i.d. matrices, thus we don't intend to provide a complete survey in the theory of random Schrodinger operators. Moreover we are aware of that a lot of pioneer and connected works are not cited since we have focused our attention to a precise mathematical aspect of the subject.
CHAPTER I
THE DETERMINISTIC SCHRODINGER OPERATOR
I.1 The difference equation. Hyperbolic structures
be the linear space of complex sequences T = (Tn) where n
Let
runs through the set of integers FL . The operator H is associated to two
given real sequences a and b with b
n
# 0
Vn c 2Z, and acts on o+by the
formula
(HT )n = bn
1 [T1
- Tn+l + anTJ
For a complex number A every solution of the difference equation HT= A lies in a two dimensional subspace of.ice spanned by the solutions p(A)
and q(A) constructed from the initial values po(A) = q-1(A) = 1, p-1(A) = qo(A) = 0, such that
:
Tn(A) = pn(A) To(a) + qn(A) T-1(a) From now in order to avoid too complicated notations we don't write the variable A in the solutions of the difference equation. A solution Y' of
the difference equation is constructed from initial values To and W-1 by a product of "transfer matrices" Yn defined by
Y
=
ra
n IL
n - Ab n 1
-1
0
187
:
188
n S=YYn-1 n
if n >, 0
Yo
Sn = Yn1 Yn+1 ...
Y
Thus
=S
n+1 J
-1
if if n
oI
n
if n >, 0
n
,
T_
n-1
1
lyn
=S
o
if
n
-1
The transfer matrices Yn and therefore the products So belong to the group SL(2,O) of two by two matrices with complex entries and of determinant one. If A is real then Y. and Sn belong to the subgroup SL(2,]R) with real entries. The construction of the solutions of the difference equation by such products of matrices is the essential link between the two parts of this book.
of SL(2,O) is said "hyperbolic" if the eigenvalues of
A matrix 0
have distinct moduli. If V is an eigenvector associated to the eigenvalue of modulus stricly less than one, then the sequence IlanVll
goes to zero when n goes to + - and for any vector W non proportional
to V the sequence TJII goes to +- when n goes to + -. (The same situation occurs when n goes to -
,
replacing V by the eigenvector
associated to the eigenvalue of modulus strictly greater than one). Let now (sn ) n e 7l ,
that s
n
be a sequence of matrices of SL(2,¢). We say
has an hyperbolic structure in the positive direction if there
exists a non zero vector V of Q2 (called the contractive vector) such = 0. It is readily seen
that lim IIsnVII
n- +-
to V then lim Is
n->+W
WII =+ W
n
that for any W non proportional
(from the determinant property). We define in
the same way an hyperbolic structure in the negative direction. The existence (for some A) of hyperbolic structures for the sequence Sn is the most important question discussed in the following chapters. In particular for a real A the existence of hyperbolic in both directions with the same contractive vector n implies the existence of a real non zero solution Y' of the difference structures for S
lim
equation with
4'
Inl _+W
n
= 0. The next "Osseledec's theorem" is a basic
tool to prove the existence of hyperbolic structures
PROPOSITION 1.1
Let (C.), h
ri
E
i
be a sequence of matrices of SL(2,O)
such that :
(i) lim 1 LogjjC n++. n
n
-1
..
1 II
=Y
189
(ii) n lim-
Log
0
n
Then there exists a non zero vector V such that
lim
n Log
n,+-
n ... 1
VI
I
and for any vector W non proportional to V
lim
...
Log 11 C'
n
1
WII =Y
The proof of this proposition can be found in [49]. The euclidian norm of a matrix of SL(2,¢) is no less than one, thus y is non negative.
When y is strictly positive 0sseledec's theorem asserts the existence of an hyperbolic structure for the sequence sn moreover we see that the sequences 1IsnVII
and 9' n 'n-1 ... Q1 or IlsnWII converge to 0 or
+ - with an exponential rate.
We now give two obvious
(but useful) lemmas related to the
solutions of the difference equation. The Wronskian of $
..yy
and T in obis
the sequence W ( ' ) Wn(T'$)
= '4 $n =
LEMMA 1.2
$n+1 Tn
Let $ and T be two solutions of the difference equation
Hu = Au, then W(T,$) is constant and this constant is zero if and only if $ and T are proportional.
Proof
in
$n
:
S
n
for n < -1
and
N-1 $n-1J ITn+1
$n+1 S
Tn
$n
n
for n,0
Since the determinant of Sn is equal to one the result is obvious.
LEMMA 1.3
Let $ and T be in £, m and n two integers with m S n, then
we have the "Green formula" n E
bk ((HT) k $k - Tk(H$)k) = Wm-1(T,$) - Wn(T,$)
k=m Proof
:
Straightforward computation.
190
Exercise 1.4
:
Let P be a solution of HP = AP, m and n two integers with
0 < n. Prove that if bk is positive for m . k
m < -1
ITn Yn+lI +
Tm Pm-ll =
n then
2
ty
I3 m iI (boI'YoI
+ b-1IT-1I2)
1.2 Self-adjointness of H. Spectral properties
The Green formula of the lemma 1.3 suggests that H should be a
symmetric operator on a suitable Hilbert subspace of ,. From now we assume in all this chapter that the sequence b is strictly positive and bounded away from zero (that is the sequence b-1 is bounded)
We denote by YO the subspace of ou defined by
{P ce/ E bn
ITnI2 < +m}
(when the symbol E has no index of summation it's understood that this index should run through ZL). Endowed with the scalar product E b
subset of
d is an Hilbert space. If D is a dense linear n 4n Tn such that H(D) C `% we denote by (H,D) the linear operator
with domain D. Let Do be the dense linear subset of % of sequences with a finite number of terms different from zero. From the Green formula we know that (H,Do) is a symmetric operator and it's not difficult to see that its adjoint (H,Do)
is equal to (H,D1) where
D1 = {P E 4U/ HY e & _ (P e :%/ E a2 bn1 lPn12 < +m } LEMMA 2.1
If the sequence a is bounded then H is a bounded self adjoint
operator.
Proof
Let a,
HY II = (E bn-1 then by the (E
and
R2 b-1
n
be the bounds of the sequences a, b-1 and P e dam:
pn-1 + Pn-1 - an triangle inequality
Pn12)1/2 :
I2)1/2 + (E b1IY 12)1/2 + (E b-1 a2 IP I2)i/2 n n n n-1 n IT n+1
:
IIHPII. II Y11 (2a + aR) Thus H is a bounded operator on'. The assumption on the sequence b
191
implies that ki G Q2 (2Z) and hence
lim
Inl
= 0 for P e r . By the
P
---
Green formula we see that H is a symmetric operator on
and hence
self adjoint. In many applications the sequence
LEMMA 2.2
is not bounded, so we prove :
a
The operator (H,D1) is self adjoint.
Proof : As remarked in the lemma 2.1, H is a symmetric operator on D1, z
hence (H,D1) is a symmetric extension of (H,Do) with (H,D1) _ (H,Do) the result follows from the theory of self adjoint extensions of symmetric operators 19 1.
From now, when we speak of the operator H, it will be always understood that we actually deal with the self adjoint operator (H,D1). As we shall see in the exercise 2.3 below, the sequence b has not to be bounded away from zero in order that (H,D0) possesses a self adjoint extension, but such an assumption seems reasonable since a solution of HP = AP which is non zero at infinity does not have any physical meaning in general.
Exercise 2.3 (
For n e N let C (b
) 1/2
b
n
be the sequence 1/2
If we assume that E Cn = + n=0 prove that the operator (H,D0) is essentially self adjoint (and (H,D1) Cn = min (bn bn+1
'
-n -n-1)
)
is its self adjoint extension). Hint
:
Use the result of 1.4 to prove that when
a # 0 the equation
HP = AP has no non zero solution in dV2 (see [4] VII theorem 1.3).
We can apply to H the general theory of self adjoint operators
which we may find in [9]. There exists a resolution of the identity E which associates to each borel subset d of 1R the projection Ed of d& and for P, P in QID,oO,T(d) = <Ed4,P> is a bounded complex radon measure on JR
(We denote the positive measurelo
number with
m A
0, R1 = (H - XI)
-
,
a complex
by
01
is a bounded operator on '% and
the spectral theorem asserts that < RO ,P> =
J(t-A)- I
do can be written as the orthogonal sum of three closed subspaces invariant under H (Lebesgue decomposition of p) defined by
:
192
you
r
{ c
db/ °
is absolutely continuous}
16P = {$ c
Y./ °
is pure point}
dl' =
V
e
=
p / °
is singular continuous}
The spectrum Q of H is the union of
G-a,
(]-p, s which are the
spectra of the restrictions of H to the invariant subspaces
Apa, 6 p
A=
We say that the spectrum of H is pure point if we have
%s
in
pp
this case 49 has an orthogonal basis of eigenvectors. In our case this Lebesgue decomposition of 'X can be obtained by the decomposition of a single measure on ]R as it is shown below. Let ek be the orthogonal basis
defined by (ek)n = bnl do (we remark that
of
II ek 112
P e Ob, = 4`n). We denote by °m,n the measure
=
bkl and for
°em,en , by on the
measure an n and a the "spectral measure" a = °° + a-1. For each n in 7L,
pn(X) and qn(A) defined in the first section are polynomial functions in A and thus the operators pn (H) and q (H) are well defined. n
pn (H) e° + qn (H) e 1 = en
PROPOSITION 2.4
The equality is obvious for n = 0 and n = -1 and is proved by -ek+l - ek-1 + ak ek induction for all n from the identity bk H ek = Proof
:
COROLLARY 2.5
For each T in QU the positive bounded measure a
is
absolutely continuous with respect to a. Proof
:
Let 4 be a Borel subset of ]R with °(d) = 0. Then
IIEAe°1I2 = Go (A) = 0 and IIEe 1112 = a-1(d) = 0, thus by the proposition 2.4
:
E,en = pn(H) Ede° + qn(H) E,e 1 = 0. This implies that Ed = 0 and hence
aT(A) = IIEAT112 = 0.
This corollary has a direct consequence for the Lebesgue decomposition of
.
Let aa, °p, as be the three parts of the Lebesgue
decomposition of a.
PROPOSITION 2.6
With the above notations,
, q a,
(T p, (3-s are the
topological supports of the positive bounded measures
°,
°a,
°p,
moreover x is an eigenvaZue of H if and only if °(a) > 0. Proof
:
By Weyl's criterion we know that A eG-if and only if there
a
193
exists a sequence ('n).30 in D1 with lim
= 0 and
11 (H - AI)'Yn11
11 Y.11 = I.
n ->+ Taking in account that for A c IR and 'Y e D1 we have the relation
II (H- AI) Y'112= 1 (t-a)2 daY (t) the result follows for W a
other parts of the spectrum we remark that if Y =
"Lebesgue decomposition" of Y e I then oa = aTa
+
y,P
and a. For the + Ys is the
ay = ors
,
We associate to H the measure valued matrix S given by
:
,
up = aYp.
a-1,0
r
a0,-1
a-1
The spectral measures a
LEMMA 2.7
can be constructed from the
m,n
spectral matrix $ by the formula IPn
pmq )
am,n
$I
qr. Proof
:
Using the expressions of en and em given by the proposition 2.4
it's enough to compute am,n(t) = <E,, em , en >.
Exercise 2.8
Prove that {EAe°, E,e 1 / A is a Borel subset of 7R} spans
a dense subspace of I. Exercice 2.9
Prove that (an + ari+l) and a have the same nul sets.
Some authors work with the operator H restricted to the half axis
N denoted by +H
.
This operator is defined on the sequences `Y = (Tn)n,0
by
(+HP)o =
bo1
(-Pl + ao'Yo)
(+HY)n = (H'Y)n if n , 1 The Hilbert space + db , the spectral measures +am,n , the resolvent +R
are defined in the same way than for H and we set +a = +a o
Exercise 2.10
Prove that
(i) pn (+H) e° = en (ii) For each Y' in
respect to
(n a IS) + ,the
(iv)
+a Y,
is absolutely continuous with
a
(iii) +am,n = Pn pm +a +
measure
o
{ EDe
(m,n a 1V)
/ 4 is a Borel subset of IIt} spans a dense subspace of
+,{
ov
194
For m,n E IN prove the following relations
Exercise 2.11
M
Pn Pm d+a =
n
-2
+
A pn pm d a = bn
(ii)
:
m do
b- 1
n
-1
an dm - bn
-1
bm
The next exercise shows that when
n+1 ({dm
n-1
+ Sm
J mA # 0 then there exists an
hyperbolic structure in the two directions for S , with distinct n contractive vectors. Let +m(A) = J(t-a)-1 d+a(t) for
Exercise 2.12
(i) We define the sequence ( P
for n
m(A) Pn + qn
Prove that H
+T
= A
n e 2Z, by
for n , 0
+'Yn = Tn =
),
m X# 0
+T
-1
+T-1
= 1
(ii) Deduce from (i) that for n 3 0
_ +m (X) pn + qn
(iii) Prove that
+ b n=0
+T.1
2
r
- J m(+m(A))
n
6ma +
Hint
RA -
(iv) The same property holds for the negative half axis and a complex number A with
m a# 0 is not an eigenvalue of H. This provesthe above
assertion about hyperbolic structures. Exercise 2.13
Prove that it's possible to recover the sequences (an)n>,0
and (bn)n>,0 from +a.
Hint
use 2.11.
:
Exercise 2.14
The sequences +T(a), -'V(A) are constructed for the two
half axis in the same way than in 2.12. Prove that the symmetric matrix
= a
m X O 0) is given for n < m by
(
+`1(m)
`Y (n)
W( T, +y')
(where W is the wronskian)
Exercise 2.15 that n k=0
2
Let A be a real number and n a positive integer. Prove d
d
bk Pk = Pn a Pn+l - pn+1 dl pn
195
1.3 Slowly increasing generalized eigenfunctions
We say that Y'
is a generalized eigenfunction if H'Y = AY for a
E
real A and 'Y is non zero. (When ' E
'Y
is an eigenfunction). A fairly
general theorem about Carleman operators asserts that such sequences
are slowly increasing [4]. Actually it's possible in the simpler case of difference operators to give a direct and easy proof of this result. As we have seen in 2.5, °m
,n
is absolutely continuous with respect to °
(thank's to the inequality I°m,nl ` 2 (Qm + °n)). Let S be a matrix of
density of S with respect to a.
For a almost all real A, the matrix S(A) is symmetric and
LEMMA 3.1 positive.
Proof
:
For each borel subset S of ]R, the real matrix $(5) is symmetric,
positive since ao,-1 = °-1,0' and for real x,y
(x,y) $ (4) xy1 1= 0x e0 + y e1 (5) The conclusion follows easily.
>, 0.
\\1
For a almost all real A there exists a generalized
PROPOSITION 3.2
eigenfunction Y' such that for each E > 0
-T (n) lim InI ±m Ii -+E
=0 ra
For ° almost all X we can write
Proof
O1
K* where K is an
S = K ILO S
orthogonal matrix, a and 6 are two positive numbers with a , S, a +a =1. By virtue of the lemma 2.7 0
nnn =
( P,
4 )K
Pn
[0
SJ
K ( qn
°
/'Y
If we denote by
see that on >, a
the firstcolumnof K and 'n
°
= pn'o +
we
bn' and that the
2 `Yn 0 . We know that on(IR) =
0
qn'Y-1
f2
1+E
sequence bnlis bounded then for any e > 0 we have
E
n#0 n
n
E
n#0 n
,f2 thus
(
< + oo
1+E
o a.e. 'P2
0
This implies that lim
InI >m
n
1+E
° a.e.
do < + m
196
The same result for the operator
Exercise 3.3
H on the half positive
2
axis
for +a almost all A, lim
:
Pn
n->+- n l+e
= 0
Prove that the Proposition 3.2 could be
Exercise 3.4
For a almost
:
all real A there exists a generalized eigenfunction 'Y such that for every
c
we have
t2(TL)
c 22(7L) .
1.4 Approximations of the spectral measures
The spectral measures a are generally obtained as weak limits m,n of spectral measures of the operator H restricted to "boxes". These
limits are independent from the boundary conditions on the boxes and this allows us the construct new approximations which are absolutely continuous with respect to the Lebesgue measure on]R.
A box A is a finite subset [M , N] of 7L , and we say that a sequence of boxes is going to 7L if the associated sequences of M and N are going
to - - and + - respectively.
For a box A anda real number x we define the operator n Hx on the Hilbert space of sequences ('Yn) , n e [M , N], (with the scalar ,gyp
product inherited from 7) by (AHxY)M = bMl (-i'1
(AHx'Y)n = (HT
N=
1 + aM'M)
for M+l
)n
bN1(-T
n
N-1
N-1 + (aN x) Y'N)
Let ^p and "q be the solutions of HW = A'Y with initial values
^qM = 0. It's easy to see that Hx is a self FPM qM-1 = 1' PM-1 adjoint matrix on ^90, its eigenvalues are the roots of the polynom '
^ PN+1
nam X
they are simple and if A is an eigenvalue we have
(A) _
N
APn
Apm(kEM bk Pk)
Let x c1R, m,n a 7L be fixed, then the sequence
PROPOSITION 4.1
converges weakly to a
Proof
:
m,n when A goes to 7l.
Fix A with 3m A
0, x c ]R, n e 7l . A function
n
4'
iA
be extended to
by ^'n = 0 if n 0 A ; with this convention
in
n ax m,n
nt At can
197
( (nH-a
I)n'Y)
(H-aI)nY'
n'YM
+
e
M-1
+
n
TN eN+l
- xn Y' N eN
When
kl - + the sequence ek converges weakly to zero, thus if the Al is bounded when A goes to 2Z (and hence weakly compact) the sequence eM-1 + ATNeN+l sequence AO = ATM x A T eN converges weakly to zero N n A 4'
is bounded by
RA en the sequence
when A goes to 7L . Taking A T =
b-lJIm al-1 . With this choice we have en = (H-A)AP + A and
n
R. e n =
A Y
RA"
If T is any weak cluster point of AT we have
.
'Y = R. en and thus A R. en converges weakly to % en when A goes to TL
Taking in account that I(t-A)_1
_
=
d
AOx
n(t)
(t-A)-1 d om,n(t)
I
we obtain the expected result since it is well known that convergence of integrals of the continuous bounded functions t -i
t1X
for mA # 0
implies weak convergence.
LEMMA 4.2
For
m X
Aq eM>_- n N+1
= aIITII2- 2iylTN12
thus
(3
mA)
IT I
'
'
N12 and this is impossible, the right member of
'N
this equation being strictly positive. We recall without proof some basic facts about the Poisson kernel of the upper half plane. These results are proven for the unit disc in [32]. Let C. be the Cauchy distribution with density 1 II
(t-a)S2+S 2
dt,
z = a + i6 with S > 0. Let f (z) be an analytic function on m z > 0 such that f can be continuously extended to IR U{m}. Then this extension satisfies the Poisson formula
f(z) =
f(t) dCz(t)
:
(Sm z > 0)
J
Furthermore let F(z) be the function associated on m z > 0 to a positive ua+us+up is the bounded measure p by :F(z) _ J(t-z)-1 dp(t). If u =
Lebesgue decomposition of p, then (i) lim 6m F(x+ic) exists for Lebesgue almost all x of IR and is equal t 0
to the density of
a.
(ii) (µs+pP) is supported by the subset {x a IR/lim 5m F(x+iE) _ +m} E+0 In particular if lim im F (x+it) exists and is finite for all real x, Ey0 then p is absolutely continuous with respect to the Lebesgue measure on
IR and its density is given by this limit. Its also useful to remark (
that
dC
J Z dp(t), that is,
m F(z)
m F(z)
is the Poisson integral of
dt
p and the above statements are nothing else than Fatou's theorems. %
We define Aam n as the average of A am n with respect to the Cauchy measure e by the formula A
aM'. (A) =
1am,n(A) dCi(x)
PROPOSITION 4.4
The measure
am
n
has a density with respect to the Apnnpm
Lebesque measure on IR given by
2 II
A2
pN+1+ pN
199
din
In particular we have the .formula 1
Proof
:
A 2 PN +
1
II
nax
Thank's to the relation
Ap np Aax n m M
=
m,n
=
1
2
PN+1
it's enough to
prove 4.4 for m = n = M. By the lemma 4.3, for 3 m i> 0 the function z
nqN+1
z -* -
qN
is continuous on Ci m z3 0 and has obviously a limit _ ZAPN pN+l when IzI -> + m . From the lemma 4.2 and the properties of the Poisson integral
nq
- iA q
N
N+1
(
F(A) = II1 J(t-A)-1 do(t) M
A
A
pN+1 - 1 pN {{
As it is readily seen that for A real limb m F(A + it) _ 64.0
1
A 2 pN+l
+ n Pi2
the result follows from the above remarks.
The sequence
nconverges weakly to a by the proposition 4.1 m,n m,n (this is true for am,n and hence for the averaged Cr
when A goes to 7L
measures). One of our essential goals is to obtain (under some hypothesis) that a
is pure point, and it seems surprising to work m'n n x am which is absolutely continuous (rather than with am,n n which is pure point). But we shall see in the chapter III that this nti
with
choice is well adapted to the proof of. localization in the random case.
Exercise 4.5 (HT)n
Hint
Compute
^
and am,n for the Laplace operator am,n
-fin-1 - Tn+l
pn(A) _ sh(n-M+1)w
A
(where ch w = - 2
sh w d am,n
and for A c ]-2,+2[,
da
Prescribing
Exercise 4.6
cos[(m-n) Arcos - 2 (1-A2) -1 /2 211
boundary values at each endpoint of the box AHy'x
A, we define for x,y a JR the operator
by
(nHy'x 1')M = bMl(-'FM+1 + (aM y)TM)
if M+1 : n < N-1
(AHY,x `F )n = (HT )n
(^Hy'xf)n =
bN1(-TN-l
Prove that for' m A # 0
- xApN+y(AgN+1 - x'qN)
200
Let +Hy be the operator on the "box" 10, +-[ with the
Exercise 4.7
boundary condition y at the endpoint
0 (see 4.6).(Wh+n y = 0 this is
just the operator +H defined in 2.10. Let +m(A) = J
dto(t)
defined in
2.12.
m(A)
(i) Prove that =
1-y+m(A)
X
Hint
use 4.6 with N = 0, x = 0
:
(ii) Prove that the measure
+ay
++'a'
= J
dCi(y) is absolutely continuous
with respect to the Lebesgue measure on R. + MM
Hint : J m
m(a)
is bounded on J m A> 0
1 - i+m(A)
(Remark that Exercise 4.8
mA
> 0 m +m(A) > 0) .
Compute
m(a) for the Laplace operator (exercises 2.12
and 4.5).
1.5 The pure point spectrum. A criterion.
Let A be an eigenvalue of H associated to a normalized eigenfunction f (the eigenspaces are one dimensional !), then
a m,n (X) = 0, there exists for]P almost all w, an hyperbolic structure in the two directions of a. If we think to theip a.e pure point spectrum, we see that we must obtain the a.e in the reverse order, namely
:
for
8 almost all w there exist hyperbolic structures in the two directions of Tl with the same contractive vector for o(w) almost all A. This
inversion of a.e properties is the most important problem discussed in the sequel.
As we are mainly concerned in this book with the independent case we give without proof a result of S. Kotani [35] asserting that y is "often positive". We say that the process (a
b n) is deterministic if n (ao,bo) is measurable with respect to the subsigma algebras generated '
by {(ak,bk)/k -11.
PROPOSITION 3.3
Assume that {A e]R/ y (X) = 0} is of positive Lebesgue
measure, then the process (an,bn) is deterministic.
The proof given by B. Simon [58] assume that (a) is bounded, but the general case is handled in [33]. A typical example of deterministic
211
process is given by the Mathieu operator and the next proposition shows that the converse of 3.3 is not true.
The very easy proof given here is due to to M. Herman [27].
PROPOSITION 3.4
Let H(w) be the Mathieu operator
:
[H(s) Y].
-Yn-1 - Yn+1 + S T. dos(211 .(% + w)
where a,s
are real numbers with a irrational.
Then for any complex nwnber A , y(a) >. Log2I-
For a complex number z we set z = z e2IIin a and we define the n matrix Zn by Proof
:
2 (zn + 1) - Azn Zn(z)
-Zn
= z
0
n
Observing that Z(e) = znYn(w,A) we have
...
IIZn (els)
and thus
Z.(eiw)II
= IISn(s,X)II
:
(211
Yn(X) =7E LogjISn(s,A)II = 1 J The function z -> Log1IZn(z) ... Yn(X)
LogJIZn(eis)
...
Zo(eiw)II dw
0
211
Zo(z)II
is subharmonic and this implies:
LogIIZn(0) ... Zo(0)II = n LogJ2I
Exercise 3.5
Compute the Lyapunov exponent for the operator
-fin-1 - Pn+1 + (1)n Tn and compare spectrum of H computed in 2.6 (Hf)n
{A/y(A) = 0} with the
.
11.4 The Lyapunov exponent in the independent case
In this section it is assumed that we are in the independent case and that1E Log(1+Ia! +Ibi) < + - the random variable (b) taking any real value
In the independent case the Lyapunov exponent is given by the
212
Furstenberg's formula involving an "invariant measure" and this allows us to prove extra properties of positivity or continuity for this exponent.
We recall there some definitions given in the first part of the book. For a non zero vector x of ]R2, x is the corresponding point of
the compact projective line denoted by X. A matrix g of SL(2,]R) acts on X by g.x = g and for two probability measures li, v on SL(2,]R) and
X respectively, we define the "convolution" p p * V(f) = J
v
by
f(g.x) du(g) dv(x) where f is a continuous function on X.
We say that V is p invariant if p 'F v = v. For a real A
,
p
a
is the
law of the matrix Y
I
=
a-Ab
-1
Let v be a probability measure on X such that
.
0
1
v(xo) = 0 for
xo=`O (11
ti
then we can define the image v of v on R by the u
application x -*
v
where x =
1v
We shall say in the sequel that we are in the "Helmotz case" when there exists a value a'F of A and a constant c such that the law of (a,b)
is supported by the line a -X*b = c. In this case the operator H can be
written
:
(HT )n = bn1[-fin-1 - 'yn+l + c'n] + a
and the matrix Y has the form Y
=
I
L
c-(A-a'F) b
-1
:
1
OJ
1
PROPOSITION 4.1
If the law of (a,b) is not concentrated on a single
point of ]R2 then for A E n (A # a'F (i)
'Fn
in the
Helmotz case. :
There exists a unique p1 invariant probability measure v, , and
VX(X) = O, vX E X.
(ii) Y M > 0 and y (X) = J JLog Imo- dpA (g) dv1 (x) (iii) If there exists a > 0 such that IE(lafa + lbla) < +- then Y(A) = ILog Itl dv1(t)
213
Proof :Looking at the theorem A.II.4.4 its enough in order to prove (i) and (ii), to verify that the closed subgroup G. of SL(2,]R)
generated by the support of uX leaves no probability measure fixed on X
r1
Under the hypothesis of 4.1 G. contains at least two matrices
gI =
g2 = and
IL
0
rl
Letting d = a-S we see
with a
JI
1
0
that G. contains the matrices u = g1 g21 = _
1
0
61
0
11 and
1
v = gl g2 = l
Ithen uxo = xo
If we set xo
.
d
1
-
1
and for every
0 /
x of X the sequence un x converges to x
Thus any fixed probability o measure under the action of G. must be concentrated on xo, but .
v.xo # xo and such a probability measure does not exist. The hypothesis of (iii) implies that IIgIIa is 1i1 integrable and we know from the theorem A. VI .4 . 3 that there exists
> 0 such that
sup Jd-S(x,y) dvx(x) < + - where d is the usual distance on X defined yEX
f .
IIIXIInIIyII
d(x,y)
by
Choosing y
2 S'2
2
IuZI
l
r
dv; (x) < +- and hence
v
I
1
=(11we obtain for x =I\v/ u f 2
2
Log (u Z ) dvx (x) < +v
J
ti v Let now m and t be the positive measures on X such that m and k are the Cauchy and Lebesgue measures on R. -1 -1 I_-II2 dg m dg lm (x) = = We have the relation dg R dm IgxII
dg
-1 R
.
di
dk
(x-)
(x)
dm
2
Letting f(x) = dR(x) = II
the above result implies that Log f(x)
2v 2 u +v
dg-lm
is vA integrable and remarking that -1 (x) = f(g.x) we obtain dg
-1
rr
II(Log
+
og a-) dud dv Log-4k
I
111111
t
= 0
dm
dg- R
Applying now Furstenberg's formula
Y(A)
(( 2 JJ
_-2 REMARK 4.2
r J
di-1
-1
m dud dvX
Log
2
dm 2
Log (v-2) dvX (x) = u
(
f
11 Log
Log I tI
dua dv1
di
dvAti(t)
The exponent y is easily computed in the Helmotz case at
the point A* as the spectral radius of the matrix
214
rc
1
1
0
and we find
:
L
if Icl < 2
y(A*) = 0 y(a*) = Log
2(IcI +
cV
`-4)
IcI 3 2
if
In the general ergodic case we have seen that y is
uppersemicontinuous (3.1). In the independent case y is actually continuous and we divide the proof in two parts (which is easy) and the case A = A
the case A # a*
be a sequence of real numbers distincts from as and
Let A
LEMMA 4.3
:
(which is more involved).
with limit A. Then for any subsequence extracted from vX
and weakly n
convergent to v the function y(A) has a limit along this subsequence equal to
:
I(ao,v) =
JJ
Log
iLg2cL
dud (g)
IIXII
°
Moreover v is uA
d\) (K)
invariant. 0
Proof
:
f (X) =
Define for each A E R the function I
Log
du (g). From ILog
I
X
. Log IIgII, it is readily
1
1
IIXII
a-ab
IIXiI
11 0 J then
seen that fX is continuous on X. If we set g, = C 1
Sup If X1 (x) -
fX2 (-X) I
IE Logllg, gall)
, JR
XEX
2
1
This yields the first result writing
Y(an) - I(Xo,v) = % (fX- fA 0
n
n
)
:
+ (VAAn - va o
)
fa
o
In order to see that V is uXinvariant we note that uX4 vA n
o
and taking in account that
= vX n
converges weakly to uXo on SL(2
'JR.)
n
and
pXn that X is compact the proof is complete.
COROLLARY 4.4
The function y(A) is continuous at each point ao
distinct from a4. Proof
Let an be a sequence with limit a0 # a
know that each limit point of vA n
From the lemma 4.3 we
invariant, hence equal to vA
is uX o
.
0
215
and y(on) converges to
converges weakly to v.
This implies that v.
n
o
I(ao,VA ) which is equal to y(ao) by 4.1. 0
In the Helmotz case y is continuous at the point A
COROLLARY 4.5
(Assuming that (b) is bounded when Icl Proof
:
> 2).
We know from 3.1 that y is an uppersemicontinuous non negative
function, hence each point A with y(A) = 0 is a continuity point of y and this gives the result for Icl . 2 in view of 4.2. In the hyperbolic c
1
1
O
> 2, the matrix
case Icl
has two real eigenvalues el and e2 with
and we denote by x1 1) the associated > 1 > l e2 l el x2 = C = eigenvectors. Each uX invariant probability measure v is a convex
(1)
I
l
combination of 6_
x
Thus continuity
and 6of y at
,
and we see from 4.2 that I(a6_ ) = y(a*).
,
the point A
follows from 4.3 if we prove that
with An -> A* is such that v(x2) = 0.
any limit point v of a sequence vA n
Let T be the closed subset of X defined by T = {x/u , v 3 O or u`< v < O} . It is readily seen that T is invariant under the action of any matrix
-1 1
Cd
with d > 2. Assuming c > 2 and the variable (b) bounded, T is rc-(A-X x)b
0
-ll
for A near A
invariant under the action of IC
A near a *
,
.
Thus for
0
1
JII
vA is supported by T and any limit point V is also supported
by T. Since we have el > 1 > e2 > 0 we see that x2
T. The same works
for c < -2 replacing v by -v in the definition of T. REMARK 4.6
(i) Corollaries 4.4 and 4.5 are direct consequences of
general theorems about perturbations of random product of matrices which
have been proven by H. Hennion [25], H. Furstenberg,.Y. Kifer [21]. In the case Icl > 2 and (b) bounded, D. Ruelle [54] has proven that y is
actually analytic at A = a*. But when lcl.= 2, analyticity can break down as shows the exercise 4.12 (iv). Such singularities have been
studied by B. Derrida [16] . (ii) In the case c = 2 (c = -2) the matrix
unique invariant probability measure 6- with
I
has an
xl
1
xl = l
with x-l =I
(6X /
-1
1
l)).
l
(iii) In the case IcI
< 2, the matrix
can have a lot 1
Cc
0
-1 1
216
v) does not
of invariant probability measures v but the integral I(A
depend on the choice of v and is equal to zero. In order to see this, we remark that such an integral does not depend on the choice of the norm on]R2 since if we set f(x) = Log
is any norm on
where IIjxIII II x
1R2we have
JJ (f(g.x) - f(x)) dv(x) duA*(g) = 0 c
1
Q 1 is a
Now there existsa real inversible matrix Q such that Q
l
0
rotation and choosing the norm IIIxjjI = IIQxII we obtain I(a*,v) = 0.
We now turn our attention to a result of E. Le Page [ 43 ] improving the corollary 4.4. We say that a function f defined on an open subset U of ]R is locally Holder continuous if for any compact T GU there exists
a strictly positive a = a(T) and a constant C(T) such that If(t) - f(s)I : C(T) It-sia
PROPOSITION 4.7
,
s,t e T
Assume that there exists a strictly positive constant s
such that]E Log(Ials + Ibis) < +- . Then the function y(A) is locally
Hd Zder continuous on JR \ {X *} . Proof of IIt2
:
We assume that the law of (a,b) is not concentrated on a point in which case the result is obvious.
The most important part of the work is to prove that the function A+ v
is, in some sens, locally Holder continuous. Using this property Furstenberg's formula will give the result for y(A). The proof uses
heavily the Holder spaces £ introduced in A.V
and the spectral
properties of the operator PXf(z) = Jf(g.R) du(g) on £ , for a small enough. Our essential goal is obtain some uniformity (with respect to A in a compact set) of the spectral decomposition of the operator PA.
Unfortunately this property is not easy to check since the function A - PX is in general not continuous from]R to the space of bounded
operators on £a. We give below the essential steps of the proof of 4.7. In all the lemmas,T is a fixed compact subset of ]R \ {a
LEMMA 4.8
There exists a(T) > 0 such that for a e 10, a(T)], A E T,
the spectrum of PA acting on %a is contained in a disc of radius strictly less than one, except the eigenvalue 1 which is simple.
Moreover for f r Y,a
217
f = VA(f) II + QA(f) where the spectral radius of QA is strictly less
P X
than one and Proof
:
1 (x)= 1
vx e X, vA(QAf) = 0, QX(II) = 0.
For each fixed A this is just the theorem A.V.4.3. The
uniformity with respect to A E T is a direct consequence of the uniform convergence of
]E LogjISn(A,w) XII
to y(A) on {A,%}
F-
T x X
(We remark that this result gives again the continuity of y(A) on T). n
In order to obtain this uniformity we remark that for a sequence 1 n-1 k (xn An) converging to (x0 ,a0) the sequence vn ,
converges veakly to vX
(each limit point is 0
1
1A
n JO ua 5x invariant)n Moreover
0
looking at the proof of 4.3 we obtain that -IF, LogjISn(Xn'w)XnII
converges to vA (fA 0
LEMMA 4.9
vn(fa
= y(A0) and this yields the result.
)
n
0
There exists R(T) > 0 and for a e 7 O,B(T)J a constant C(a)
such that for (s,t) e T, f E
IIPsf - Ptflla
. C(a)Ilfjja It-sla/2 /2
Proof
Using the notations of A.V.2.4 and 4.3 we have for f E ct a
s,t E T
(a >01)
:
(i) IPsf(X) - Ptf(X) I < ma (f) 1E[5a (gs.x, gt.x)] Since 5(gs.x, gtx)
IbI
IIgsiI
Is-tl we obtain from the
IIgtII
integrability condition in 4.7 the existence (for a small enough) of A(a) such that
:
IPsf(x) - Pt f(x)I . ma(f) A(a) I s-tia
(ii) IPS f(x) - P f(Y)I . ma S
5 (g Since
s
X, g
(f) 1E
5n( gs X, gs. Y) 5a (X, )')
Y) s
AN
.
Thus if
we note by NH the operator H restricted to the box [0,NJ with zero boundary conditions at each endpoint and Nk its distribution of states, we have
:
IVA,N(X-)
-JE[Nk(]_m,A
N+1
We know that vA N converges weakly to vA
,
that ]E[Nk] converges weakly
to k (by 6.4) and that v. and k are continuous measures thus
lim N->+-
va(X) = 1im ]E[ k(
k( -co,A])
N- +-
We now turn our attention to a remarkable relation between the Lyapunov exponent and the distribution of states first pointed out by D. Thouless
62] and hence called the "Thouless formula". We follow here
the proof given by J. Avron, B. Simon [3] and W. Craig, B. Simon [12].
228
PROPOSITION 6.7
Assume that the random variable a is bounded and
E(Log b) < +- , then for each complex number A y(A) _ ]E(Log b) +
Loglt-al dk(t) J
Proof
:
Let A be a complex number with 3 m A # 0 and N 1,l. The roots of
the polynom pN+l
(qN+1), are the eigenvalues of the operator H '
restricted to the box [O,N]
endpoint. Denoting by Nk
([1,N] ) with zero boundary condition at each
(Nk) the distribution of these eigenvalues it
is readily seen that N
r
Loglt-al dNk(t) + 1
J LogIt-Al
N+1 kEO Log bk
dNk(t) + N kE1 Log bk =
N+1
LogIgN+1(X)I
N LogIgN+1(A)I
The random variable (a) is bounded, thus the probability measures Nk, Nk are supported by the same compact set. The Birkhoff ergodic theorem
and 6.4 mplie that
Nm N+1
Log
:
pN+1(a)I
N
+
N LogIgN+1(X)I
= JLoglt- Al dk(t) +]E(Log b),]P a.e.
We have thus proved that each entry of the matrix S
= N
rpN+1 pN
qN+111
qN JI
has
the same asymptotic behaviour. Since we know from 3.2 that
lim N-+- N
LoglISnII = y(A)
we can conclude that for
]P a.e. 6m X0 0
y(X) _ ]E(Log b) + J Logl t-aI dk(t) Each function in the above formula is a subharmonic function on
for
y this is the proposition 3.1 and the subharmonicity of JLoglt-al dk(t) is straightforward. Since two subharmonic functions almost everywhere equal, with respect to the Lebesgue measure on ¢, are everywhere equal, we obtain the result.
REMARK 6.8
:
We have assumed that the random variable (a) is bounded in
the proof of the Thouless formula. But if we only assume that
]E Log(1 + lal +b) < +- then y(X)
]E Log b and JLogl t-al dk(t) are well defined (the latter for 6 m A 0 0 by 6.5). This suggests that such a ,
229
formula should be true under this weaker hypothesis...
We know from 6.3 that the distribution function K(A) = k(]-m,Xl) is continuous. We list below sow extra regularity properties of K. K is said locally Log Holder continuous if for all compact subset T of R there exists CT > 0 such that for s,t e T, Is-tI
we have
IK(s) - K(t)I < CT (Logls-tl-1)-1
PROPOSITION 6.9 (i) In the general ergodic case assume that (a) and (b) are bounded, then K is locally Log Holder continuous. (ii) In the independent case assume that (a) is bounded and that there
exists a > 0 with]E(ba) < + - then K is locally HOZder continuous on
R\A
.
(iii) In the independent case assume that bn = 1, Vn a 7L , ]E(a2) < +
and that the law of (a) has a density on R, then K has a continuous derivative.
(iv) In the independent case, assume that the conditional law of (a) given (b) has a bounded density $b on R, such that]E(bjkkbll) < +
Then
k has a bounded density with respect to the Lebesgue measure. Proof
(i) Since (b) is bounded there exists S such that Sb < 1 and then by the Thouless formula
:
Y(SA) -]E Log Eb =
Loglt-Al dk(t)
v where K(t) = K(lt).
Thus this latter integral is positive for all X e 0
and this implies that K and hence K must be locally Log Holder continuous by standart analysis arguments [12]. (ii) The result follows from 4.7 and the properties of the Hilbert transform [47]. For, P
E:
L2(]R)
we denote by 'T its Hilbert transform and
if ]A,B[ contains the support of k, we choose P(A) = K(A) if A e ]A,B[, T(A) = 0 if A i ]A,B[. By integration by parts and the Thouless formula: ti
1
(Y(A) - LogIB-AI) for almost all A with respect to the Lebesgue 4'(A) = II measure. Letting O(A)
(y(A) - LogIB-AI)
,
@ is locally Holder continuous on
IR%{B,A*}, hence 0 is also locally Holder continuous and 0 = T
a.e
hence 0 = P on R \{B,A }. Since B can be chosen outside of any compact
230
we obtain the result.
(iii) This result will be proven in 111.4.1.
Let a be a number with )m a> 0, N an integer. We denote by
(iv)
of the operator H restricted to the half axis [N, + m [ or
the resolvent
]-
N
,NI
Following the lines of 1.2.12 and 1.2.14 we define
Nm (A) =
± n = N
N. The functions +'Y can be extended to the whole of
where n > N or n
7L in such a way that these extensions satisfy H'Y = A'Y, and we find N' _ (Nm)(Np)
+ (Nq)
where Np and Nq are solutions of H'V = AY' with the
,
boundary conditions N +pN
N +q
_ N
_ N
-pN
N-1
1
-qN+l
N
N
N
N
+'N-1
+qN
-pN+l
-qN = 0
Then it's easy to verify that
< RA e N ,
:
-1
eN >
1
(. + m) +
N-1 m) + AbN
(
- aN
The functions (N++m) and (N-1 m) are functions of the process (an,bn) for n >. N+1, n < N-1 respectively and have positive imaginary parts. Taking expectation in the above formula, first integrating with respect to
(aN,bN), we obtain
f JELb
This yields the expected result, using Lebesgue's convergence theorem in the formula
k(h)
_ o II
:
h(x) ( im f
dk(t) ) dx
J
where h is a continuous function with compact support.
Let L1'a be the space of functions f e L1(IR)
such that there
a
exists g e L1(IR) with f(t)(1+t2) /2 = g(t)
(where f is the Fourier
transform of f). B. Simon and M. Taylor [61] have recently proven that in the independent case with bn = 1, n e ZZ, if we assume that (a) is bounded with a density in L1'a for a strictly positive a ,
then K is a
231
function. A typical example is given by the uniform distribution on
C
a compact interval.
When the density of (a) has stronger regularity, namely when (eita)l . lIE
a e sltl for some positive constants a and a such that
a < 8, then K has an analytic continuation to a strip {z c
/l) mzl 0. This result is proven by R. Carmona in [91 using the "Molcanov formula". Further results about Lifschitz exponents can also be found in this course.
Exercise 6.10
In the general ergodic case assume that (a) is bounded. Nk ' Prove that the sequence of probability measures (N , 1) converges
weakly almost surely to k where dNk _
II(N+1)
N
1
1
da
P2 + P2 N
2
k=O bk Pk
N+ 1
The two following exercises show that the distribution of states can be the same in two models which have in general very different spectral properties ([59] and [18j).
Exercise 6.11
"The Maryland model", see [18]. Let H = H + V.
Laplace operator HT = -yn-1
- 'n+1
is the
and for a positive number 8 and
irrational a we define (VT)n = 8 Tn tg(IIn a + a) where
a e [O,IIJ \ {II/2 - IIn a}n e 2Z = Q. This is not exactly a quasi-periodic model since a is not bounded but a lot of explicit computations can be performed in this "solvable model". Let U be the operator (U'Y)n = 2 iw
e2iII n aT
n
and z be the complex number
e
(i) Prove that H = H + i8 (zU - I) (zU + I)
(ii) Prove that for
mX>0
(zU+I)(I+z CU)-1 R1+i6 where C = Rx+i8 (H - (A - i8) I ) (iii) Observing that
1Cll < 1 expand (I+z C U)
in power series and
prove that lE [R A]
= Ra+is
Conclude that k has a density and k = k '
Ci6 (Cis is the Cauchy law
232
and see I.4.5 for k). Hint
dk as
:
for A real. dk(t)
lim 1. c+o
J
Exercise 6.12
t-(A+ir)
dk(t)
=3m 1
t-A-i6
The "Lloyd model", see [60]. Let H = H + V as in the
exercise 6.11 but now (VP)n = a
n
Y'n where a
n is a sequence of real
independent random variables with a common law Ci6 (see 4.12) (i) Let I be a finite set of integers (perhaps not all distinct) and s
k
= 1.
k c I, a set of positive numbers with E s
keI k
Prove that IE [exp it E sk ak] = exp(-sltl) keI
(ii) "Dyson's formula". Let A and B be two squares matrices with complex entries and t a real number.
Prove that et(A+B) =
etB + ft es(A+B)A e(t-s)B °
Iterating this relation, expand e
t(A+B)
in a convergent series.
(iii) Applying (i) and (ii) to the operators prove that
and
A
H and A H
esltl eit H
I[.eit
F
ds
HI
= e -St eit H
(iv) Conclude from (iii) that k = k
i6
11.7 The pure point spectrum. Kotani's criterion
We assume in this section that (b) is strictly positive, bounded away from zero and that IE Log(1 + lal
+ b) < +
Let a +(a -) be the subsigma algebra of a generated by the
process (a n b n) for n >,N (n,<M) and integers 11,N. For a '
measurable positive function
defined on
we denote by
+i the conditional expectation with respect to a-+ and for each
Borel subset d of the real line we set a+(4) = IE [a(A) l a+] . a+(0) are a+ measurable functions defined IP almost everywhere. It's possible to construct a topology on 52 such that D becomes a Polish space
233
and a_ its Borel sigma algebra. Thus there exists a choice of a+(A) such that a+ are positive measures on ]R with a+(]R) = o(IR) .
By a
monotone class argument we see that if F+ are measurable functions on 0 x IR with respect to the product of
with the Borel sigma algebra
of IR then
IE IJ( F+ do I a+Q
F+ da+
=(
and
J
IE If F+ do] = IE If F+ do+] (take care that a+ have nothing to do with -a of 1.2.10) We are now able to prove Kotani's criterion [36].
Let A be a Borel subset of IR and m a positive
PROPOSITION 7.1.
measure on A. We assume that y(X) > 0 absolutely
m.a.e on A and that a+ are
continuous on A with respect to m,
Then for ]P almost all w the exponent
almost everywhere
IP almost everywhere.
y(A) is strictly positive a(w)
on A, the spectrum of H(w) is pure point on A (if
any) and the eigenfunctions decay with the exponential rate -y(A). Proof
:
Define the subsets W+ of 0 x IR by
w e 0, A E A, lim
InI Log IIYnII = 0
n->±°
w± = { (w,X)
/
lim 1 LogilsnII = y(A) > 0 n->±m
InI
a+ are absolutely continuous with respect to m
The sets W+ are measurable with respect to the products of a+ by the Borel sigma algebra of IR and we know from the ergodic theorem 3.2 and the hypothesis that W+ are of
full IP ®m measure.
Let F+ be the functions defined by F+(w,A) = 1
if
(w,A)
= 0
if
(w,A) e W.
W+
Using the preceding discussion and the absolute continuity of a+ with respect to m
IE If F+ do] = IE If F+ do+] = 0 Then by Fubini's theorem, for IP almost all w, for a(w) almost all A in A, each solution of H'1 = AP is, in each direction of 2Z, growing or
decaying exponentially fast (thank's to Osseledec's theorem I.1.1). Looking at the proposition 1.3.2, for o(w)
all A there exists a slowly
234
increasing solution of HP = AY' and thus this solution must be
exponentially decaying in each direction of a with rate -y(A). From the integrability condition Log(1+b) < +- we know that for IP almost all w,
lbn(w)I . n for n large enough. This implies that a(w) is
supported by the eigenvalues of H(w), hence a(w) is pure point and the eigenfunctions decay with the rate -y(X).
REMARK 7.2
The above matching method of two solutions of the
difference equation was employed by S.A. Molchanov L45] and
R. Carmona Is I . If the process (an,bn) is deterministic then the hypothesis of 7.1 does not hold because in this case a+ = a and we know from 5.2 that if y > 0
m
a.e. on A then for IP almost all w, a(w) is orthogonal to m.
Thus in order to check the hypothesis of 7.1 some randomness is inevitable. We shall see in III sufficient conditions in order to deal with criterion 7.1.
11.8 Asymptotic properties of the conductance in the disordered wire
In physical litterature there exist several approaches to the notion of conductance. We only give there a formal definition and we refer for instance to J.L. Pichard [51] for physical background. We call "plane waves" two sequences u and v of the form u = einB n vn = e in0 where e is a real number different from a multiple of R. Then any sequence Y of complex numbers can be "projected" on the plane waves u and v according to
n+1 Y
where T n Rn
It
=
vn+l
+ 8
n un
n
a 1n
u.+1
- a
n vn
un+l
vn+l
u
v
n
n >, 0, from
= T
an
n 8n compute
is an invertible matrix. In order to
n
a-1 8-1
we set Z n = T n l Y n Tn-1
,
Ti
n = Zn n
n-1
'
Z
o
is readily seen that Rn = Ti nSn T_1 and thus if Y< satisfies the
235
difference equation (HT)n = XT relation
n
, n c [0,
..
,
N], we obtain the
:
=
B=1
11N
\
I
1
a
This means that if we "input" the wave of "coordinates" ( the disordered box A = 10, (coordinates
..
,
1)
into
NJ, then the "output" wave have
SN
aN The reflexion and transmission coefficients tN and r,, of the disordered box A = L0,
0
..
/
HN
,
NJ are defined by the equation
(N / 2
t
and the conductance of A is the ratio C,, = IrN I
Our problem is to
find the assymptotic behavior of CN when N -
Let S U (1,1) be the subgroup of matrices in SL(2,Q) of the form
Ia B
Tn1
a
Y
). It's an easy exercise to verify that if Y e SL(2,IR) then is in S U(1,1). Thus ZN and therefore RN are in S U (1,1). As
//Tn-1
a consequence we may write
CN = I
:
(
2
(IIlNxIl2 - 1) where x is the vector I
I nNXII =
T_ I
I
1
)Of .2. Since
T_
nl Sn T-1xl I
=
I I
nl Sn y II
where y = T_lx, the asymptotic
behavior of CN is obtained from the asymptotic behavior of the columns of SN which is well known in the independent case.
THEOREM 8.1
Assume that the sequence (a
n
b
n),
n >, 0, is an i.i.d
'
sequence with values in TR2 and that its law is not supported by a single point of IR2. Then for A c IR (A # a* in the Helmotz case) (i) The sequence 1 Log C converges IP a.e. to -2y(A) when n - + n n Moreover y(A) is strictly positive.
(ii)Assume that there exists a > 0 with IE(Iala + Ibla) < +
then
there exists a strictly positive number a2(A) such that the law of the
sequence N (Log C law with
Vill'
ariance
n
+ 2n y(X)) converges weakly to a centered Gaussian
a(X). 2
236
Proof
I
:
From the inequalities
I Tn I I - 1
:
IISnYII 1 IITn1 Sny II : IITn111 IISnYII
we can write Log I Tn1 Sn Y I = Log I I Sn y I + Un where the random sequence Un is bounded since I Tn II . 2, 11 Tn1 II . I sine I I
I
I
Thus we obtain V n
1n Log C n =
n
-
?n
Log IISnYII
where
Vn = [Log2 -2Log Un - Log (1 - (IISnYII + Un)-1)1 We know from 4.1 that y(A) > 0 and from A.II.3.6 that Log Its yII V converges IP a.e. to y(X). This implies that n converges IP a.e. to n n zero and the result (i) follows. In order to prove (ii) we write
1 (Log Cn + 2n y(A))
V
:
V
2
A
(Log Its nYII
- n y(A)) +
n
/
'
The result (ii) now is a direct application of A.V.4.3 and of the IP
a.e. convergence of °n
to zero.
REMARKS
(i) The same result was already obtained by A.J. O'Connor [48] and T. Verheggen [63] under stronger assumptions on the law of (a,b). Actually they suppose that this law has a density with respect to the Lebesgue measure on IR. (ii) Extensions to the strip, using results of chapter III and A.IV.
can be obtained in a similar way and this is what is done by R. Jhonston and H. Kunz in I30].
CHAPTER III
THE PURE POINT SPECTRUM
We now prove the strongest singularity property for the spectrum of Schrodinger operators, that is the pure point spectrum property. Until the end of the year 1984 the only known proofs were very close from the original one given by the Russian school (see the introduction). In most of these works it was not clear what was really necessary in order to obtain the pure point spectrum property, since they dealed essentially with the independent or Markov cases under strong assumptions on these processes. A very clarifying idea of Kotani introduced in 11.7 and applied to the independent case in the first section gives very easily the expected result, at least inthe classical Schrodinger case, that is when bn = 1, Vn e 2z. It's not clear that such a procedure can be easily applied to the Helmotz case. Furthermore Kotani's theory does not use any approximation of the spectral properties of H by means of "thermodynamic limits" along sequences of boxes going to 2Z. Thus if we want to obtain some information about the speed of convergence of these spectral measures computed in boxes, we have to use an other way. In this purpose the second section is devoted to the study of the Laplace transform on SL(2,IR) . The corresponding
operators were already introduced in A.V. In section 4, using the spectral properties of these operators,we prove again the pure point spectrum property in the independent case, thank'sto the "good" approximations of the spectral measures of H obtained in I. These results are also used in section 5, in order to obtain a representation
237
238
formula for the density of states.
III.1 The pure point spectrum, first proof.
In this section we suppose that (an'b ) is a random process defined a n on some complete probability space (O,¢.,1P) but we don't require
ergodicity or stationarity. We only assume that for IP almost w (i) The sequence bn(w) is strictly positive and bounded away from zero.
(ii) lim 1 Log(1+lan(w)l + bn(w)) = 0 n-*±
Ini
This last assumption is used to check the hypothesis (ii) in Osseledec's theorem I.1.1 and to establish that for 0 < r < 1 the series E b rn is n 8 a.e convergent.
Carefull readers notice that the proof in 11.7.1 has nothing to do ,with ergodicity if we assume the existence of y properly. Thus we can prove exactly in the same way as in 11. 7. 1
PROPOSITION 1.1
:
Let A be a BoreZ subset of IR and m a positive measure
on A. We assume tnat there exist IP x m
a.e
lim 1 Log lSnl = y+
strictly positive functions
Furthermore we a.e. n+± nl suppose that a+ (defined in 11.7) are absolutely continuous with respect y,(w,A) such that
IP O m
tom on A. Tnen for IP almost all w, y+(w,X) are strictly positive for a(w) almost
all A, the spectrum of H(w) is pure point on A (if any) and the eigenfunctions decay with the exponential rate -y+(w,X)
As remarked by S. Kotani [36] and F. Delyon, Y. Levy, B. Souillard [14], if we suppose that the process (an,bn) is an
independent sequence of random variables with values in JR2 (not
necessarily identically distributed), then the assumption on a+ is easily checked
COROLLARY 1.2
Let (an,bn) be an independent process such that there
239
two functions y+(a),
exist
each A, lim 1 Log1S n InI
I
m
= y+
a.e strictly positive and that for a.e. (where in is the Lebesgue
IP
n-*±
measure on IR). Furthermore we suppose that there exist
two consecutive indexes n =no,
no+1, such that the conditional law of an given bn has a bounded
density b with IE (I I b II ) < + n
n
Then the conclusions of 1.1 hold, replacing a by an 0
Proof
:
+ an
+1'
0
In view of proposition 1.1 and exercise 1.2.9 we have only to
prove that for n = no and n = no+1
IE[a Ia +7 is absolutely continuous n with respect to the Lebesgue measure m. Choosing N = n0+2 and M = n0-1
in the definition of
in 11.7 these conditional expectations are
nothing else than expectations with respect to the random variables (an,bn) for n = n0 and n
we know that for
= no+l. Looking at the proof of 11.6.9 (iv)
m A > 0 and any n
:
_ -L(n++m) + (n-lm) + abn - anj where
n+1
MM is a function of (ak,bk) for k . n+l,
n-1m
and
(A) is a function of (ak,bk) for k . n-1,
with strictly positive imaginary parts. Denoting by IE
n
the expectation with respect to the single random
variable (abn) we obtain r
rlEn (on)] (dt)
m
III
%
n
t-a
Thus IEn (an) has a bounded density with respect to the Lebesgue measure on IR and the result follows.
Note that for each A the limit of 1 LogflSo1I is IP a.e. constant and
Ini thus we can choose y+ as functions of A only.
COROLLARY 1.3
Let (an,bn) be an independent identically distributed
random process such that IE Log (1 + IaI + b) < + = and that the law of
(a) given (b) has a bounded density
b
with IE E II b I I ] < + = . Then the
conclusions of 1.1 noZd. Proof
:
From
11.4.1 we know that y(A) > 0
everywhere (the assumption
on the law of (a) excludes the Helmotz case).
Thus the ergodic theorem 11.3.2 and corollary 1.2 yield the result
240
(Remark that the hypothesis (ii) of the begining of this section is
satisfied since IE Log (1 + al + b) < +m). REMARK 1.4
(i) In the Helmotz case, in example for the operator b- 1
(Hi')n =
n
(-fin-1 - 'n+l),
we obtain in the proof of 1.2 1
(n++m) + (n-1m-) + abn11 j
we don't obtain n the absolute continuity of c1+ in general. (See however the remark of and if we integrate with respect to the variable b
S. KOTANI in the appendix).
(ii) In the independent case, it would be very tempting to choose as the tail sigma algebras generated by the process (an,bn). Thus o+ would be obtained as expectations rather than as conditional expectations . Unfortunately it's far from evidence that the
mesurability property for the sets W+ introduced in the proof of 11.7.1 can be easily checked....
111.2 The Laplace transform on SL(2,IR)
Let G be the unimodular group SL(2,IR) , X the one dimensional real
projective space, C(X) the Banach space of complex continuous functions on X endowed with the supremum norm. As explained in 11.4, G acts
continuously on X if we use the usual norm topology for the two by two
matrices and the distance 6 on X defined by 6 (x, y) =
JJ
x n Y11
since it is readily
,
seen that
llYll
11x11
6(gl.x ,g2.y) . 6(5,Y)
11g211
I1g111
+ I1g1-g211 I1g211
Moreover a sequence xn converges to xo in X iff there exists a sequence
gn in G such that
x = g n.xo. n converges to the identity in G and n
g
In order to check this last property we observe that for
(rcos0
x
\r sin 8)
,
r e IR \ {0}
,
0 e [O,n [ we can write x = g0. x- where g0
is the rotation of angle 0 and x, = (1).
241
Let now u be a probability measure on G and t a real number. We define the operator Tt by T
t f (x) = J (--x-,-) f (g.x) du (g)
,
f e C(X).
IIgxII
It is readily seen that if f Ilglln
du(g) < -- for n > O,then Tt is a
positive bounded operator on C(X) for t c [-n,+n] and the Markovian operator To is just the transition operator for the left random walk of law u on X. We denote by Tt the dual operator, acting on the space M(X) of complex measures on X by (Tt v)(f) = v(Ttf)
,
v e M(X), f e C(X). A
probability measure v such that T+ v = v is called a p invariant 0
probability measure, such measures always exist
from the Markov-
Kakutani theorem (see A.I.3.5) but are not unique in general. Most of the sequel is devoted to the study of spectral properties of the operators Tt. These operators have already been introduced in A.V but we have to notice that we use them on C(X) rather than on Holder spaces. (see remark 2.10.(ii)).
Let n be a strictly positive number such that
PROPOSITION 2.1
J IIgIIn du(g) < + - . Then the map t -> Tt is analytic from [-n,+nJ to the space of bounded operator on C(X). More precisely if
Dt f (x) = J
then
(Log) IIgx11
nn I( tt-
)
t f (g.x) du (g)
IIgxII
- DS)f(x)I: IIfII
(IITt+EII+IITt-EII+IITS+EII+IITS-EII) 2E
S
for f e C(X), n >,
1,
e > 0,
(s,t) a
L-n,+n]
.
Proof : We first remark that the n th power of the above operators are
the operators associated with the n th power of convolution un of u on SL(2,7R) and that
du (g)] n
JIIgJJ' dun(g)
[ JIIgIIf
Log r(t) is convex
on [0,2]. From 2.2 this function is strictly negative at a point in 0,2[ and hence is strictly negative on ]0,2[.
(i) Most of the results of this section are true in a much
REMARK 2.10
more general setting. Actually it's possible to traduce all of them to the case of the action of a locally compact group 0 acting continuously on a compact metric space X. If we replace
Log
11x11
by an additive
llgxll cocycle a introduced in A.I.3.2, the Laplace transform is given by
Tt f(x) =
ot(g,x)
dp(g)
J
See for instance [41] and chapter IV. (ii) In view of A.V.2.5 the operator To acting on
possesses the
same decomposition as in theorem 2.9, without any density assumption on p. By perturbation theory such a decomposition holds for Tt if t is small enough (A.V.4.3). But we need a decomposition of T2 and this is the reason why we don't use the Holder space Za and work with C(X).
Exercise 2.11
It's not always true that v has X as support, even when
p has a density. (Think of p supported by the matrices of SL(2,1R) with non negative entries, then v is supported by the invariant set X+ = {x / x = (UV) / uv >, 0}). Suppose that p is supported by the
matrices of the form L1
where t is a real random variable such
0]
that the topological support
of the law of t contains an open set of
-2, +2[. We know from the proof of 11.4.1 that v is unique. Prove that the support of v is equal to X.
Exercise 2.12
Without any assumption on p prove that T1 is a bounded
operator of L2(m) and is self adjoint if p = u Denoting by y the limit of the subadditive sequence JLog
llgll
n
j
Log Ilgll dpn(g) (assuming that
dp(g) < + m) prove that the logarithm of the spectral radius
of T1 acting on L2(m) is greater than -y.
247
Hint
1 Log II Tn I n 1
I
> 1n Log f
IIJl du
n
- 1n 1 Log I I g II dun (g)
(g) dm(x)
IIgxII
As a consequence if we know that this spectral radius is strictly less is not amenable, see Y. Guivarch
than one (this is the case when G u
and Y. Derrienic [171) then y is strictly positive.
Suppose that u is supported by the matrices
Exercise 2.13
where t is a real random variable distributed
Lt
1 0J according to the Cauchy
v
law Ci6 with
> 0. Compute the eigenfunctions p and 'v of T2 and T2
with m(p) = m(p) = 1.
111.3 The pure point spectrum, second proof.
In order to obtain the pure point spectrum property in the general independent equidistributed case,we apply the results of the preceding section together with the criterion given in 1.5.1.
We denote by uA the law of the matrix
Y =
ralba
ll
1
Tt,A the operator
Tt associated to the law u, and by v. a ua invariant probability measure.
PROPOSITION 3.1
Assume that IE (a2+b2) < + -. Let J be a compact subset
of IR and suppose that for A e J, the law of a - bA has a density T with respect to the Lebesgue measure of ]R such that A -> TX is continuous from J to LI(IR) . Then there exists positive constants AJ,
BJ, CJ with AJ < 1 and such that
(i)
(ii)
IITI X11 I
5 BJ AJ
I T2A II
Proof
:
CJ
,
,
n e IN, A e J
n e IN, X e J
An easy Jacobian computation shows that u3 has a density
with respect to the Haar measure
of G and that A ->
X is continuous
from J to L1(dg). Then following the lines of 2.6 it's readily seen that
A -> Tt
1
is continuous from J to the space of bounded operators
on C(X) for t e [0,2]. We now observe that it's enough to prove (i) and
(ii) for Tt
rather for Tt'A since X
IITt,AII : J IIgIIt dua = IE([2 + (a-Ab)23t/2)
248
Looking at 2.9 we know that the spectral radius of T3
is strictly
less than one and that T2 X(f) = m(f) pX + Qf(f) where pA is the
continuous density with respect to m of v , and the spectral radius of QX is strictly less than one.
The continuous perturbation theory of compact operators (see Kato [31]) asserts that the maping A --> QX is continuous from J to the space of bounded operators on C(X) and A --- pA is continuous from J to C(X). The
conclusion of 3.1 now follows from the next lemma
:
LEMMA 3.2 Let A -TX be a continuous mapping from a compact set J to the spaceof bounded operators of a Banach space, such that for
each A e J the
spectral radius of TA is strictly less than 1. Then there exists AJ , no
,
A e J,
IIT'II
AJ
The Logarithm of the spectral radius r(A) of Tx is an
uppersemicontinuous function which attains its maximum value f on the compact set J, thus t < 0. Furthermore i = Sup(Z, Log r(A)) is the pointwise limit of the decreasing sequence Sup(f, inf k LogIIT,'II) of
1fk,
(iii) If X and Y are real symmetric matrices of order Az with Y non
negative then (^PN+1 - (X + iY)'PN) is invertible for J mA>0 (Take care that actuaZZy^p N is a function of A
!).
We now want to compute the mean value of ox with respect to a A m,n probability measure on the symmetric matrices X analogous to the Cauchy measure in the one dimensional case. For this we recall some facts about the Poisson formula of the upper Siegel half plane (see [28]).
257
For k = 0,1,
..
k, we set Sk = {X + iY/X and Y are real symmetric
,
matrices of order Z with Y non negative and of rank k}. Then there exists a family P(Z,dX) of probability measures on So
,
indexed by
Z e L) Sk called the "Poisson kernel". If f is an harmonic function k=1 (see [281 for the definition) in each Sk, k = 1, ..., R , which can be continuously extended to S
0
U {W} we have the representation theorem R
f(Z) = 1
f(X) P(Z,dX)
,
Z E U Sk k=1
PROPOSITION 1.5
The measure valuated matrix no
m ,n
=
I
1
wax
m,n
P(iI,dX)
nas a density with respect to the Lebesgue measure on 1R given by Ao
d om,n
= 1 Ap (At)-l R m
dl
where
AP'
n
A4 = ^pN+1 APN+1 + ^PN APN
AGMX then it's enough APn AUX,n APm to prove the above result for m = n = M. From proposition 1.4 we know
Proof
It's easily seen that
:
that the function LAPN+l
,
- ZN -1L^QN+l
- ZnQN] is analytic and hence
harmonic, on U Sk, and extends continuously to So U {W}. Thus from k=1
,t
the above discussion we obtain (d m l > 0 ) d
nti
j
(t)
=
_ LnPN+l - 1APN] EAQN+l - i^QN]
t
n
nP N+1
Taking in account that the matrix A P
N
that for each real A r
lim ' m
e-o since
QN+1
is symplectic, we see
AQN
:
d"am(t)
f t-(A+it)
=
(AA )
We remark that the matrix nt is invertible
:
PN+1 nPN
II
^QN+1
= YN YN-1 ...
A
QN
YM I0 L
0 II J
Following 1.5.1 we now establish a sufficient criterion implying
258
the pure point spectrum property for the operator H. Let S(A) be matrix a
of density of
0,o
a-1,ol
with respect to a. If A is an eigenvalue
Loo,-1 Q 1, of H the terms of S(A) are well defined and let T1,
..
,
'fr an
orthonormal basis of the associated eigenspace. We denote by To, (T-1) <ei,o,
the R x r matrix with entries
Ti,
,
(<e i,-1, Tj>)
It is readily
seen that ,T
S(A) =
1
(T', T'
I
To
GM T
l
The matrix( T
\
)
1
0
1
I
is of rank r thus S is also of rank r. Furthermore
1
from the relation S' J Sn = J and if we denote by Ln the 2R x r matrix
n
I1
,f2
+1
1+1
,
fr Tr+1I
we obtain Ln J Ln = Lo J Lo
.
Since the
n
n
eigenfunctions are going to zero at infinity, Lo ' J La = 0 and this
implies that the rank r of L
is no greater than R. Finally we obtain
0
that the rank of S(X) is equal to the rank r , R of the eigenspace associated to A and that each column of S(A) can be taken as initial values of an eigenvector.
THEOREM 1.6
Let I be an open subset of IR for which there exists two
positive constants C and p with p (m,n) E A, i,j e [I, ...,R
^ti I
(j,n)I
a
(I)
Then if EIIBnjj1/2 p1n1
C P
±W
InI
p ; 2.
for 1
(The limits for n = t- are equal since IISnII = IISn1II)
The subadditive ergodic theorem implies IP almost everywhere convergence
of 1 LogjiApSnII to 11(A) + ... yp(A) (for a fixed A). Moreover the I nnl
integrability condition IE Lo5(1 + II All + IIBII) implies the IP almost everywhere convergence of 1 LogHIY n to zero. Thus for each fixed A, II
In j
Osseledec's theorem 1.1 applies IP
a.e..Looking at the proof of 11.3.1
we see that the sums yl(X) + ... + y (A) are subharmonic, but except p
for y1 we can't claim that the Lyapunov exponents are subharmonic. The measurability property 11.2.1 still holds and we obtain the Pastur's lemma about eigenvalues of H looking at the proof of 11.2.2. The only change is that now, the trace of the spectral projector E. on the eigenspace associated to the eigenvalue A is no greater than R. Then the Ishii's, Pastur's theorem becomes
THEOREM 2.1
:
Let A be a Borel subset of IR such that the Lyapunov
exponent y5(X) is strictly positive almost everywhere on A with respect to the Lebesgue measure. Then aa(A) = 0 ,IP.a.e. Proof
:
We define the measurable subset W of 0 x IR by
Jwe0 W
=
(w,a)
A t A, lim
I+ InI
I
LogflYn(w,A) II
=0
A is not an eigenvalue of H(w)
lim
InI- InI
LogIIAP Sn(w,X) II, = YI(A) +
for p = 1, ..
... +yp(a)
, Z
The proof is now the same than in 11.5.1 remarking that y5 > 0 implies that the two contractive subspaces V+ and V A is not an eigenvalue V+(1V
= {0). Thus
are 2 dimensional. Since
IR25
= V+ ® V
and each non
zero solution of HY = AT is growing with an exponential rate at least
in one direction of ZZ. REMARK 2.2
No converse of 2.1 is known for the strip. In the one
dimensional case Kotani's proof uses heavily the Thouless formula. Such
261
a formula has been proved in the strip by W. Craig, B. Simon [12] in the following form
:
Y1(A) + ... + Y(X) = k
Loglx-AJ dk(x) J
They assume Bn = I, Vn e 7l, and that A is a bounded matrix. But it
seems that no formula of this kind is available for the individual exponent Yf(A) since it is believed that this exponent is not subharmonic in general. A weaker form of Kotani's converse would be "Let A be a Borel subset of positive Lebesgue measure contained in {a/y1(a) = 0) then a
a
(A)
> 0,
IP
a.e.". Such a result certainly can be
proven using the above Thouless formula, the m functions introduced in exercise 1.8 and following the lines of [58].
Exercise 2.3
Ishii's proof of 2.1 (see exercise 11.5.6) extends to the
positive strip
:
(i) Prove that for 5m A > 0 Pn+1_
Pn+1)-1
+Ra e eo> =
-
n
(Pn
Qn+l
(ii) Prove the expected result, following the lines of 11.5.6
Exercise 2.4
Let Y be a real symmetric matrix of order i and +HY the
Schrodinger operator with boundary condition Y at the point zero. Prove
that for 3 m A i 0 and a box A= 10,N] Y (i) < R; eo , e o> _ - [PN+1 + QN+1 Y1 (ii)i+Ra e , eo> = M [I -YM]
(where M =
0M
1
QN+1
introduced in exercise 1.8)
(iii) Denoting by
+o Y
the trace of
+o Y
0 o prove that
+tio
= f
+oy
P(iI,dY)
has a bounded density with respect to the Lebesgue measure on IR. (Hint : M = U + iV with U and V symmetric, V positive definite).
(iv) Following the lines of exercise 11.5.7 prove the generalization of Kotani's result
:
"Let A be a Borel subset of IR such that 11(X) is strictly positive almost everywhere on A. Then for IP almost all w, for almost all Y (with respect to the natural Lebesgue measure R
d yi,3 on the
i>,j
symmetric matrices) the spectrum of +HY is pure point on A (if any) and the eigenfunctions are exponentially decaying".
262
Hint
:
P(iI, dY) is absolutely continuous with respect to II d y, l.J i%J
(see [28]).
We now adapt Kotani's criterion 11.7.1 to the strip
PROPOSITION 2.5
:
Let A be a Borel subset of IR and m a positive measure
on A. We assume that yt(A) > 0 m.a.e on A and that a+ are absolutely continuous on A with respect to m,IP a.e. Then for IP almost all w,
y1(X) is strictly positive for a(w) almost all A, the spectrum of H(w) is pure point on A (if any) and the eigenfunctions fall off exponentially with a rate given by a Lyapunov exponent.
Following the lines of 11.7.1 we obtain that for IP almost all w, for a(w) almost all A the contractive f dimensional subspaces V+ and V have a common non zero vector. But we don't know in which Osseledec's subspace of V+ and V
lies this common
vector, neither if this common
vector is unique. Furthermore given such a common vector the rate of decay of the associated eigenfunction has no reason to be the same in the two directions of 2Z
!
From a physical point of view it seem's
natural to think that this common vector has to be unique and that it does not belong to proper Osseledec's subspaces of V+ and V-. Thus the rate of decay would be given by -yk(A) in the two directions of 2Z, but we don't know any proof of this ....
IV.3 Lyapunov exponents in the independent case. The pure point spectrum (first proof).
As seen in section 2, positivity of yR is a crucial hypothesis to obtain absence of absolutely continuous spectrum or pure point spectrum.
In contrast with the one dimensional case the only known proofs of this positivity require strong assumptions on the law of (A,B). In this section we suppose that (An,Brl) is an i.i.d. sequence of random
variables. In order to avoid too complicated proofs, we also assume that (A) is a Jacobi matrix, (B) a diagonal matrix such that the I l of (A,B) are independent random variables with random entries (a,b) l
263
values in uR2 . We say that the "Helmotz case" occurs when, for some
such that
index i, there exists a constant cl and a a value ai
ai -A *i
b
i
=c
i
Moreover we suppose that each (b1) is positive, bounded away from zero
and that lE Log (1 + IIAII + IIBII) < +m . We denote by ua the law of the 0J1
-aB LAI
matrix
L
Assume that no (al,bl) is supported by a single point
PROPOSITION 3.1
(Assuming A ¢ X
of IR2 . Then y1(A) > 0
Proof
:
if the Helmotz case occurs).
Following 11.4.1 we see that the subgroup G. of SP(f,IR)
generated by the support of u, contains matrices of the form u = 10 v = LA
0]
I]
where 4 can be chosen diagonal with no nul diagonal entries.
El
projective subspace spanned by the I first 1' (E2 ) the 22 2R x we of IR (last) basis vectors of IR Then for a vector x 11
We denote by
ll
jx +n4x \
obtain unx
I` x2
2J
x
vnx =
x2
X1 2
+ n d xl
If x e E 1 then If x ¢ E1 then u
n
-
f 6x21
. x converges to ` O /which belongs to E1. This
implies that each fixed probability measure on P(IR
2k)
under the action
of GA is supported by E1. The same is true for E2 and as a consequence GX leaves no fixed probability measure on P(IR21)
thus the result
follows from A.III.7.5.
We have thus obtained under broad conditions positivity of y1. As seen in section 2 we actually need positivity of yR and this is certainly the case if all the Lyapunov exponents are distinct. Unfortunately we have to work with contractive properties on the exterior powers of IR2R ua
,
of the semi group generated by the support of
as explained in A.IV. Such a property can hardly be directly
checked for Schrodinger matrices and we are led to prescribe stronger assumptions the law of (A,B) in order to make use of a more tractable criterion given in A.IV. Let F be the application of IRS'
EA(x) F(x)
L
I
in G = SP(I,IR) defined by
-I 0
where A(x) is the Jacobi matrix whose diagonal
elements are given by the coordinates of the vector x. In the one
264
dimensional case (1 = 1) a straightforward computation proves that the F(x3) F(x2) F(x1) of IR3 in G is regular at
application
with x2 40 (d is then of rank 3, which is the
each point of IR
dimension When I
of the Lie algebra L(. of G).
2, such a computation becomes tremendous thus we have to work
>,
in a more intrinsic way. This is what is done in [40] and we only give here a sketch of the proofs, for readers familiar with Lie algebras.
Let U be an open set in IRQ, then the closed subgroup GU
LEMMA 3.2
generated by Proof
:
F(U) is equal to G.
As in the proof of 3.1 we see that GU contains matrices of the
form u = 1I
and v = L I
IJ
where A is a diagonal matrix whose
diagonalsr runs through an open set of IRQ. We can write u = exp 10
01
and thus by varying A we can construct a family of
v = explA 00]
(J
elements of the Lie algebra U of C. For R.= 1 this family generates "]gas a Lie algebra and the proof is complete. For R. > 1 we have to verify that conjugates of these elements with respect to some matrices
of F(U) generate { . Let U be an open set in IR(X1, ..
LEMMA 3 . 3
elements of rank of
Then there exists (al,
the family (where gk
{Ad(g11) Xk
,
=
Ad(g11g21) Xk,
..
F(ak))
,
,
Xr) some non zero
an ) in U
such that the
0
:
Ad(g11g21,
..., gnl) Xk / k= 1, ...,r} 0
is equal to the dimension of Proof
:
Consider the set of subspaces of `
families for all the finite subsets (al,
..
spanned by the above ,
an) of U. It is easily
seen that a subspace of maximum dimension in this set is invariant under Ad g-1 for g in
F(U)
and hence is an ideal of e{, from 3.2. 1 i
a simple Lie algebra thus this maximal subspace must be equal to
For each open set U in IR1
LEMMA 3.4 set (al, (D(xl'x2' (al,
..
there exists an integer no and a
..., an ) in U such that the map ..
,
, an)
xR ) = F(xn) ... F(x2) F(x1) is a submersion near 0
0
0
Proof
:
Choose as (X,,
10 ..
XR)
matrices of the form
0
A.1 pJ where
265
A. has zero entries except the term (i,i) which is equal to one. Let (al,
an ) be the subset of U given by lemma 3.3. For t cIRR and
..
o
R
F(a+t) = (exp T) F(a) and thus
t.X. we obtain
T = I
i 1
i=1
5(a1 +tl, ..
,
+ to
an 0
(exp Tn ) F (an ) ... exp(T1) F (a1) 0
0
0
The result follows from 3.3 and the computation of the differential of S at the point (al,
..
,
an ). 0
REMARK 3.5
It's certainly possible to choose explicitely (al,
and to see that n
= 2R +1, but the computation is tremendous
o
2R+1 is obvious since dim
lower bound n
,
!
an
(The °
= R(2R+1).
We now formulate the expected results of this algebraic manipulation
THEOREM 3.6
(i) Suppose that for i = 1,
..
,
R the supports of the
laws of (a.-Xb.) contain an open set of IR, then there exists an integer no such that the support of ubo contains an open set of SP (9.,IR).
(ii) Suppose that for i = 1,
..
,
R the laws of (a. -abi) have a n
density on iR, then there exists an integer no such that }1X° has a density with respect to be Haar measure of SP(R,IR). Proof
:
Let U. be an open set contained in the support of the law of i
no
R
(ai -Xbi) and U = R
Ui then by 3.3, S(U
) contains an open set.
i=1
Furthermore t is an analytical mapping and thus its Jacobian function is an analytical function which is not always zero. This implies that the zeroes of this function are contained in a countable union of lower In0
dimensional subspaces of IR
,
hence negligible with respect to the
Lebesgue measure.
COROLLARY 3 . 7
Suppose that for i =1, ..
,
z the supports of the Law of
(ai -abi) contain an open set of IR or that the Law of (ai - abi) have
a density with respect to the Lebesgue of al, then y9(A) > 0. Proof
:
direct application of 3.6 and A.IV.3.5.
Before proving the main theorem of this section we give a useful
266
deterministic lemma. If X is a symmetric real matrix of order n then each spectral measure is a function of the
n(2+1) entries of X. We
denote by ak the spectral measure associated to the vector ek of the orthonormal basis (e1,
..
,
en) and xk the diagonal term of order (k,k)
of X.
With the above notations
LEMMA 3.8
J ak(dt) dxk = II dt
Proof
It's possible to suppose k = 1, then X can be written
:
xI
X =
u'1
where u is a (n-1) dimensional vector. For m a> 0 we set
u
R
1
rT
P = (X -XI)-1el and from the decomposition P =1
1
we obtain
xIT1 + u' = A Y1 + 1 ti
T1u+X = a0
Let h = (X -XI)u ,
then
4 = -PIh and P1 = (x1 -u'h A)1 r
Taking in account that u'h =
au(dt)
J
rr aI(dt)dx1
Jm
=3M
t-1
THEOREM 3.9
f
we see that
3m(u'h) > 0, hence:
t-A
(x1 -u'h -x)- dx1 = II 1
Suppose that each random variable (a ), i = 1, i
..
,
1,
possesses a bounded density with respect to the Lebesgue measure. Then if H is the classical independent Schrodinger operator, the Lyapunov exponent yz(A) is strictly positive VA E IR. Moreover for IF almost all a the spectrum of H(w) is pure point and the eigenfunctions fall off exponentially with a rate given by a Lyapunov exponent. Proof
:
Looking at 3.7 and 2.5 we have only to prove that a+ are
absolutely continuous with respect to a fixed measure.
We choose a generated by
(An)n>l'and
box A = 1-N,N], denoting by IE(i,n)
random variable a
IE
(i,n)
(
i, n
we obtain from 3.8
a(i,n)(i,n))
II
a-_ generated by (An)n